DeepSeek-V3.1-Terminus-gguf

This is a new model from DeepSeek. Make sure you have enough ram/gpu to run. On the right of model card, you may see the size of each quantized models. The cheapest hardware to run full DeepSeek model is using apple Mac Studio, which can have 512 GB ram/ 9500 dollars, Or 256 GB/5500 dollars. Any other PC is not even close. Some dell AI PC selling at $5000 have only 4 GB or 8 GB gpu, which are generations behind.

Use the model in LM Studio

download and install LM Studio

https://lmstudio.ai/

Discover models

In the LM Studio, click "Discover" icon. "Mission Control" popup window will be displayed.

In the "Mission Control" search bar, type "ling1000T/DeepSeek-V3.1-Terminus-gguf" and check "GGUF", the model should be found.

Download the model.

You may choose quantized model.

Load the model.

Ask questions.

quantized models

Type Bits Quality Description
Q2_K 2-bit 🟥 Low Minimal footprint; only for tests
Q3_K_S 3-bit 🟧 Low “Small” variant (less accurate)
Q3_K_M 3-bit 🟧 Low–Med “Medium” variant
Q4_K_S 4-bit 🟨 Med Small, faster, slightly less quality
Q4_K_M 4-bit 🟩 Med–High “Medium” — best 4-bit balance
Q5_K_S 5-bit 🟩 High Slightly smaller than Q5_K_M
Q5_K_M 5-bit 🟩🟩 High Excellent general-purpose quant
Q8_0 8-bit 🟩🟩🟩🟩 Near-lossless baseline
Downloads last month
164
GGUF
Model size
671B params
Architecture
deepseek2
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

8-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ling1000T/DeepSeek-V3.1-Terminus-gguf

Quantized
(19)
this model