This model is a gguf q4km format of janhq/Jan-v1-2509 generated by intel/auto-round algorithm. Embedding layer and lm-head layer are fallback to 8 bits and non expert layers are fallback to 4 bits
original model:
https://huggingface.co/janhq/Jan-v1-2509

quant_methode:
https://github.com/intel/auto-round

untested...

Downloads last month
28
Safetensors
Model size
0.9B params
Tensor type
I32
BF16
F16
Inference Providers NEW
This model isn't deployed by any Inference Provider. 馃檵 Ask for provider support

Model tree for kalle07/Jan-v1-2509_128_4_autoround

Quantized
(74)
this model