DeepSeek-V3.1-Terminus-gguf
This is a new model from DeepSeek. Make sure you have enough ram/gpu to run. On the right of model card, you may see the size of each quantized models. The cheapest hardware to run full DeepSeek model is using apple Mac Studio, which can have 512 GB ram/ 9500 dollars, Or 256 GB/5500 dollars. Any other PC is not even close. Some dell AI PC selling at $5000 have only 4 GB or 8 GB gpu, which are generations behind.
Use the model in LM Studio
download and install LM Studio
Discover models
In the LM Studio, click "Discover" icon. "Mission Control" popup window will be displayed.
In the "Mission Control" search bar, type "ling1000T/DeepSeek-V3.1-Terminus-gguf" and check "GGUF", the model should be found.
Download the model.
You may choose quantized model.
Load the model.
Ask questions.
quantized models
| Type | Bits | Quality | Description |
|---|---|---|---|
| Q2_K | 2-bit | 🟥 Low | Minimal footprint; only for tests |
| Q3_K_S | 3-bit | 🟧 Low | “Small” variant (less accurate) |
| Q3_K_M | 3-bit | 🟧 Low–Med | “Medium” variant |
| Q4_K_S | 4-bit | 🟨 Med | Small, faster, slightly less quality |
| Q4_K_M | 4-bit | 🟩 Med–High | “Medium” — best 4-bit balance |
| Q5_K_S | 5-bit | 🟩 High | Slightly smaller than Q5_K_M |
| Q5_K_M | 5-bit | 🟩🟩 High | Excellent general-purpose quant |
| Q8_0 | 8-bit | 🟩🟩🟩🟩 | Near-lossless baseline |
- Downloads last month
- 164
Hardware compatibility
Log In
to view the estimation
2-bit
3-bit
4-bit
5-bit
8-bit
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support
Model tree for ling1000T/DeepSeek-V3.1-Terminus-gguf
Base model
deepseek-ai/DeepSeek-V3.1-Base
Quantized
deepseek-ai/DeepSeek-V3.1-Terminus