- Joined
- Jun 26, 2023
- Messages
- 58 (0.10/day)
Processor | 7800X3D @ Curve Optimizer: All Core: -25 |
---|---|
Motherboard | TUF Gaming B650-Plus |
Memory | 2xKSM48E40BD8KM-32HM ECC RAM (ECC enabled in BIOS) |
Video Card(s) | 4070 @ 110W |
Display(s) | SAMSUNG S95B 55" QD-OLED TV |
Power Supply | RM850x |
With the AI age being here, we need fast memory and lots of it, so we can host our favorite LLMs locally.
Even Edward Snowden is complaining.
We also need DDR6 quad channel (or something entirely new and faster) consumer desktop motherboards with up to 256 or 384GB RAM (a/my current B650 mobo supports only up to 128GB RAM), so we can self-host our favourite big (MOE) LLMs like DeepSeek-R1 (quants: e.g. 1, 2) (real DeepSeek are the ones without the "Distill" in the names) (MOE LLMs run (much) faster than dense LLM, when both have the same number of parameters, DS-R1 tested on 4.gen Epyc server at like 8 tokens/s.) or Llama-3.1 405B quants. Bigger LLMs will always be better than smaller ones (everything else being equal, and not specialized). Please, don't hold humanity back.
Even Edward Snowden is complaining.
We also need DDR6 quad channel (or something entirely new and faster) consumer desktop motherboards with up to 256 or 384GB RAM (a/my current B650 mobo supports only up to 128GB RAM), so we can self-host our favourite big (MOE) LLMs like DeepSeek-R1 (quants: e.g. 1, 2) (real DeepSeek are the ones without the "Distill" in the names) (MOE LLMs run (much) faster than dense LLM, when both have the same number of parameters, DS-R1 tested on 4.gen Epyc server at like 8 tokens/s.) or Llama-3.1 405B quants. Bigger LLMs will always be better than smaller ones (everything else being equal, and not specialized). Please, don't hold humanity back.