Albatros39
New Member
- Joined
- Jan 27, 2025
- Messages
- 1 (0.25/day)
But also some people are running the 685B model.Worth pointing out that the models that people are running on M4, Mac Minis and gaming GPUs have very little to do with the actually capable one DeepSeek is operating on its website. That is a completely different, much larger model requiring at least 700GB of VRAM.
What the market is concerned about is that such a capable model could be trained with ~5M USD worth of compute (excluding GPU costs). That doesn't mean though that putting more compute on it won't improve the results...
It gets 6 tokens/s with 1 Epyc CPU, using the 12-channel SP5 platform, with a maximum memory bandwidth 576 GB/s, per CPU.
The DDR4 platform gets <2 tokens/s, but with a few RTX3090 you can get it to usable speed.
You still need an expensive machine, but with used parts, its doable for an individual.