Model Card for Model ID

This model is a quantized version of Qwen3-32B converted to AWQ Q4 format using the mlx library for efficient inference. It retains the core capabilities of Qwen3 while optimizing for resource constraints.

Downloads last month
178
Safetensors
Model size
5.17B params
Tensor type
BF16
·
U32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for Goraint/Qwen3-32B-4bit-AWQ-MLX

Base model

Qwen/Qwen3-32B
Finetuned
(30)
this model