Add files using upload-large-folder tool
Browse files- README.md +0 -9
- config.json +16 -16
- model-00012-of-00012.safetensors +2 -2
- model.safetensors.index.json +1 -1
README.md
CHANGED
|
@@ -24,15 +24,6 @@ base_model: cerebras/GLM-4.5-Air-REAP-82B-A12B
|
|
| 24 |
|
| 25 |
# GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx
|
| 26 |
|
| 27 |
-
```bash
|
| 28 |
-
Model Perplexity
|
| 29 |
-
qx64g-hi 7.025 ± 0.063
|
| 30 |
-
qx86g-hi 7.031 ± 0.064
|
| 31 |
-
```
|
| 32 |
-
|
| 33 |
-
This is a custom Deckard(qx) formula that enhances Air-specific layers
|
| 34 |
-
|
| 35 |
-
|
| 36 |
This model [GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx](https://huggingface.co/GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx) was
|
| 37 |
converted to MLX format from [cerebras/GLM-4.5-Air-REAP-82B-A12B](https://huggingface.co/cerebras/GLM-4.5-Air-REAP-82B-A12B)
|
| 38 |
using mlx-lm version **0.28.3**.
|
|
|
|
| 24 |
|
| 25 |
# GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx
|
| 26 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 27 |
This model [GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx](https://huggingface.co/GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx) was
|
| 28 |
converted to MLX format from [cerebras/GLM-4.5-Air-REAP-82B-A12B](https://huggingface.co/cerebras/GLM-4.5-Air-REAP-82B-A12B)
|
| 29 |
using mlx-lm version **0.28.3**.
|
config.json
CHANGED
|
@@ -1827,11 +1827,11 @@
|
|
| 1827 |
},
|
| 1828 |
"model.layers.45.self_attn.q_proj": {
|
| 1829 |
"group_size": 32,
|
| 1830 |
-
"bits":
|
| 1831 |
},
|
| 1832 |
"model.layers.45.self_attn.k_proj": {
|
| 1833 |
"group_size": 32,
|
| 1834 |
-
"bits":
|
| 1835 |
},
|
| 1836 |
"model.layers.45.self_attn.v_proj": {
|
| 1837 |
"group_size": 32,
|
|
@@ -1839,7 +1839,7 @@
|
|
| 1839 |
},
|
| 1840 |
"model.layers.45.self_attn.o_proj": {
|
| 1841 |
"group_size": 32,
|
| 1842 |
-
"bits":
|
| 1843 |
},
|
| 1844 |
"model.layers.45.mlp.switch_mlp.gate_proj": {
|
| 1845 |
"group_size": 32,
|
|
@@ -1847,23 +1847,23 @@
|
|
| 1847 |
},
|
| 1848 |
"model.layers.45.mlp.switch_mlp.up_proj": {
|
| 1849 |
"group_size": 32,
|
| 1850 |
-
"bits":
|
| 1851 |
},
|
| 1852 |
"model.layers.45.mlp.switch_mlp.down_proj": {
|
| 1853 |
"group_size": 32,
|
| 1854 |
-
"bits":
|
| 1855 |
},
|
| 1856 |
"model.layers.45.mlp.shared_experts.gate_proj": {
|
| 1857 |
"group_size": 32,
|
| 1858 |
-
"bits":
|
| 1859 |
},
|
| 1860 |
"model.layers.45.mlp.shared_experts.up_proj": {
|
| 1861 |
"group_size": 32,
|
| 1862 |
-
"bits":
|
| 1863 |
},
|
| 1864 |
"model.layers.45.mlp.shared_experts.down_proj": {
|
| 1865 |
"group_size": 32,
|
| 1866 |
-
"bits":
|
| 1867 |
},
|
| 1868 |
"lm_head": {
|
| 1869 |
"group_size": 32,
|
|
@@ -3668,11 +3668,11 @@
|
|
| 3668 |
},
|
| 3669 |
"model.layers.45.self_attn.q_proj": {
|
| 3670 |
"group_size": 32,
|
| 3671 |
-
"bits":
|
| 3672 |
},
|
| 3673 |
"model.layers.45.self_attn.k_proj": {
|
| 3674 |
"group_size": 32,
|
| 3675 |
-
"bits":
|
| 3676 |
},
|
| 3677 |
"model.layers.45.self_attn.v_proj": {
|
| 3678 |
"group_size": 32,
|
|
@@ -3680,7 +3680,7 @@
|
|
| 3680 |
},
|
| 3681 |
"model.layers.45.self_attn.o_proj": {
|
| 3682 |
"group_size": 32,
|
| 3683 |
-
"bits":
|
| 3684 |
},
|
| 3685 |
"model.layers.45.mlp.switch_mlp.gate_proj": {
|
| 3686 |
"group_size": 32,
|
|
@@ -3688,23 +3688,23 @@
|
|
| 3688 |
},
|
| 3689 |
"model.layers.45.mlp.switch_mlp.up_proj": {
|
| 3690 |
"group_size": 32,
|
| 3691 |
-
"bits":
|
| 3692 |
},
|
| 3693 |
"model.layers.45.mlp.switch_mlp.down_proj": {
|
| 3694 |
"group_size": 32,
|
| 3695 |
-
"bits":
|
| 3696 |
},
|
| 3697 |
"model.layers.45.mlp.shared_experts.gate_proj": {
|
| 3698 |
"group_size": 32,
|
| 3699 |
-
"bits":
|
| 3700 |
},
|
| 3701 |
"model.layers.45.mlp.shared_experts.up_proj": {
|
| 3702 |
"group_size": 32,
|
| 3703 |
-
"bits":
|
| 3704 |
},
|
| 3705 |
"model.layers.45.mlp.shared_experts.down_proj": {
|
| 3706 |
"group_size": 32,
|
| 3707 |
-
"bits":
|
| 3708 |
},
|
| 3709 |
"lm_head": {
|
| 3710 |
"group_size": 32,
|
|
|
|
| 1827 |
},
|
| 1828 |
"model.layers.45.self_attn.q_proj": {
|
| 1829 |
"group_size": 32,
|
| 1830 |
+
"bits": 6
|
| 1831 |
},
|
| 1832 |
"model.layers.45.self_attn.k_proj": {
|
| 1833 |
"group_size": 32,
|
| 1834 |
+
"bits": 6
|
| 1835 |
},
|
| 1836 |
"model.layers.45.self_attn.v_proj": {
|
| 1837 |
"group_size": 32,
|
|
|
|
| 1839 |
},
|
| 1840 |
"model.layers.45.self_attn.o_proj": {
|
| 1841 |
"group_size": 32,
|
| 1842 |
+
"bits": 6
|
| 1843 |
},
|
| 1844 |
"model.layers.45.mlp.switch_mlp.gate_proj": {
|
| 1845 |
"group_size": 32,
|
|
|
|
| 1847 |
},
|
| 1848 |
"model.layers.45.mlp.switch_mlp.up_proj": {
|
| 1849 |
"group_size": 32,
|
| 1850 |
+
"bits": 6
|
| 1851 |
},
|
| 1852 |
"model.layers.45.mlp.switch_mlp.down_proj": {
|
| 1853 |
"group_size": 32,
|
| 1854 |
+
"bits": 6
|
| 1855 |
},
|
| 1856 |
"model.layers.45.mlp.shared_experts.gate_proj": {
|
| 1857 |
"group_size": 32,
|
| 1858 |
+
"bits": 6
|
| 1859 |
},
|
| 1860 |
"model.layers.45.mlp.shared_experts.up_proj": {
|
| 1861 |
"group_size": 32,
|
| 1862 |
+
"bits": 6
|
| 1863 |
},
|
| 1864 |
"model.layers.45.mlp.shared_experts.down_proj": {
|
| 1865 |
"group_size": 32,
|
| 1866 |
+
"bits": 6
|
| 1867 |
},
|
| 1868 |
"lm_head": {
|
| 1869 |
"group_size": 32,
|
|
|
|
| 3668 |
},
|
| 3669 |
"model.layers.45.self_attn.q_proj": {
|
| 3670 |
"group_size": 32,
|
| 3671 |
+
"bits": 6
|
| 3672 |
},
|
| 3673 |
"model.layers.45.self_attn.k_proj": {
|
| 3674 |
"group_size": 32,
|
| 3675 |
+
"bits": 6
|
| 3676 |
},
|
| 3677 |
"model.layers.45.self_attn.v_proj": {
|
| 3678 |
"group_size": 32,
|
|
|
|
| 3680 |
},
|
| 3681 |
"model.layers.45.self_attn.o_proj": {
|
| 3682 |
"group_size": 32,
|
| 3683 |
+
"bits": 6
|
| 3684 |
},
|
| 3685 |
"model.layers.45.mlp.switch_mlp.gate_proj": {
|
| 3686 |
"group_size": 32,
|
|
|
|
| 3688 |
},
|
| 3689 |
"model.layers.45.mlp.switch_mlp.up_proj": {
|
| 3690 |
"group_size": 32,
|
| 3691 |
+
"bits": 6
|
| 3692 |
},
|
| 3693 |
"model.layers.45.mlp.switch_mlp.down_proj": {
|
| 3694 |
"group_size": 32,
|
| 3695 |
+
"bits": 6
|
| 3696 |
},
|
| 3697 |
"model.layers.45.mlp.shared_experts.gate_proj": {
|
| 3698 |
"group_size": 32,
|
| 3699 |
+
"bits": 6
|
| 3700 |
},
|
| 3701 |
"model.layers.45.mlp.shared_experts.up_proj": {
|
| 3702 |
"group_size": 32,
|
| 3703 |
+
"bits": 6
|
| 3704 |
},
|
| 3705 |
"model.layers.45.mlp.shared_experts.down_proj": {
|
| 3706 |
"group_size": 32,
|
| 3707 |
+
"bits": 6
|
| 3708 |
},
|
| 3709 |
"lm_head": {
|
| 3710 |
"group_size": 32,
|
model-00012-of-00012.safetensors
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f4d3e33f99555de00ab1172ee77d428681ac67516f9855840b79977c35d438c4
|
| 3 |
+
size 2465541568
|
model.safetensors.index.json
CHANGED
|
@@ -1,6 +1,6 @@
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
-
"total_size":
|
| 4 |
"total_parameters": 81932185824
|
| 5 |
},
|
| 6 |
"weight_map": {
|
|
|
|
| 1 |
{
|
| 2 |
"metadata": {
|
| 3 |
+
"total_size": 59570946944,
|
| 4 |
"total_parameters": 81932185824
|
| 5 |
},
|
| 6 |
"weight_map": {
|