- Joined
- Mar 11, 2008
- Messages
- 1,069 (0.17/day)
- Location
- Hungary / Budapest
System Name | Kincsem |
---|---|
Processor | AMD Ryzen 9 9950X |
Motherboard | ASUS ProArt X870E-CREATOR WIFI |
Cooling | Be Quiet Dark Rock Pro 5 |
Memory | Kingston Fury KF560C32RSK2-96 (2×48GB 6GHz) |
Video Card(s) | Sapphire AMD RX 7900 XT Pulse |
Storage | Samsung 990PRO 2TB + Samsung 980PRO 2TB + FURY Renegade 2TB+ Adata 2TB + WD Ultrastar HC550 16TB |
Display(s) | Acer QHD 27"@144Hz 1ms + UHD 27"@60Hz |
Case | Cooler Master CM 690 III |
Power Supply | Seasonic 1300W 80+ Gold Prime |
Mouse | Logitech G502 Hero |
Keyboard | HyperX Alloy Elite RGB |
Software | Windows 10-64 |
Benchmark Scores | https://valid.x86.fr/9qw7iq https://valid.x86.fr/4d8n02 X570 https://www.techpowerup.com/gpuz/g46uc |
Hello Forum,
I occasionally run LLM-s locally with LM Studio.
And I mostly use
DeepSeek-R1-Distill-Qwen-14B-GGUF (Q8_0) ~42 token/s
phi-4-GGUF (Q8_0) ~42 token/s
DeepSeek-R1-Distill-Qwen-32B-GGUF (Q6_K) ~3,2 token/s
Llama-3.3-70B-Instruct-GGUF (4Q_K_M) ~1,9 token/s
Using DeepSeek for logical and math problems while Phi and Llama for language question usually.
Really curious what do you use and how fast is it on your rig.
I occasionally run LLM-s locally with LM Studio.
And I mostly use
DeepSeek-R1-Distill-Qwen-14B-GGUF (Q8_0) ~42 token/s
phi-4-GGUF (Q8_0) ~42 token/s
DeepSeek-R1-Distill-Qwen-32B-GGUF (Q6_K) ~3,2 token/s
Llama-3.3-70B-Instruct-GGUF (4Q_K_M) ~1,9 token/s
Using DeepSeek for logical and math problems while Phi and Llama for language question usually.
Really curious what do you use and how fast is it on your rig.