nightmedia commited on
Commit
fbcc39d
·
verified ·
1 Parent(s): fc9d5e8

Add files using upload-large-folder tool

Browse files
README.md CHANGED
@@ -24,15 +24,6 @@ base_model: cerebras/GLM-4.5-Air-REAP-82B-A12B
24
 
25
  # GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx
26
 
27
- ```bash
28
- Model Perplexity
29
- qx64g-hi 7.025 ± 0.063
30
- qx86g-hi 7.031 ± 0.064
31
- ```
32
-
33
- This is a custom Deckard(qx) formula that enhances Air-specific layers
34
-
35
-
36
  This model [GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx](https://huggingface.co/GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx) was
37
  converted to MLX format from [cerebras/GLM-4.5-Air-REAP-82B-A12B](https://huggingface.co/cerebras/GLM-4.5-Air-REAP-82B-A12B)
38
  using mlx-lm version **0.28.3**.
 
24
 
25
  # GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx
26
 
 
 
 
 
 
 
 
 
 
27
  This model [GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx](https://huggingface.co/GLM-4.5-Air-REAP-82B-A12B-qx64g-hi-mlx) was
28
  converted to MLX format from [cerebras/GLM-4.5-Air-REAP-82B-A12B](https://huggingface.co/cerebras/GLM-4.5-Air-REAP-82B-A12B)
29
  using mlx-lm version **0.28.3**.
config.json CHANGED
@@ -1827,11 +1827,11 @@
1827
  },
1828
  "model.layers.45.self_attn.q_proj": {
1829
  "group_size": 32,
1830
- "bits": 4
1831
  },
1832
  "model.layers.45.self_attn.k_proj": {
1833
  "group_size": 32,
1834
- "bits": 4
1835
  },
1836
  "model.layers.45.self_attn.v_proj": {
1837
  "group_size": 32,
@@ -1839,7 +1839,7 @@
1839
  },
1840
  "model.layers.45.self_attn.o_proj": {
1841
  "group_size": 32,
1842
- "bits": 4
1843
  },
1844
  "model.layers.45.mlp.switch_mlp.gate_proj": {
1845
  "group_size": 32,
@@ -1847,23 +1847,23 @@
1847
  },
1848
  "model.layers.45.mlp.switch_mlp.up_proj": {
1849
  "group_size": 32,
1850
- "bits": 4
1851
  },
1852
  "model.layers.45.mlp.switch_mlp.down_proj": {
1853
  "group_size": 32,
1854
- "bits": 4
1855
  },
1856
  "model.layers.45.mlp.shared_experts.gate_proj": {
1857
  "group_size": 32,
1858
- "bits": 4
1859
  },
1860
  "model.layers.45.mlp.shared_experts.up_proj": {
1861
  "group_size": 32,
1862
- "bits": 4
1863
  },
1864
  "model.layers.45.mlp.shared_experts.down_proj": {
1865
  "group_size": 32,
1866
- "bits": 4
1867
  },
1868
  "lm_head": {
1869
  "group_size": 32,
@@ -3668,11 +3668,11 @@
3668
  },
3669
  "model.layers.45.self_attn.q_proj": {
3670
  "group_size": 32,
3671
- "bits": 4
3672
  },
3673
  "model.layers.45.self_attn.k_proj": {
3674
  "group_size": 32,
3675
- "bits": 4
3676
  },
3677
  "model.layers.45.self_attn.v_proj": {
3678
  "group_size": 32,
@@ -3680,7 +3680,7 @@
3680
  },
3681
  "model.layers.45.self_attn.o_proj": {
3682
  "group_size": 32,
3683
- "bits": 4
3684
  },
3685
  "model.layers.45.mlp.switch_mlp.gate_proj": {
3686
  "group_size": 32,
@@ -3688,23 +3688,23 @@
3688
  },
3689
  "model.layers.45.mlp.switch_mlp.up_proj": {
3690
  "group_size": 32,
3691
- "bits": 4
3692
  },
3693
  "model.layers.45.mlp.switch_mlp.down_proj": {
3694
  "group_size": 32,
3695
- "bits": 4
3696
  },
3697
  "model.layers.45.mlp.shared_experts.gate_proj": {
3698
  "group_size": 32,
3699
- "bits": 4
3700
  },
3701
  "model.layers.45.mlp.shared_experts.up_proj": {
3702
  "group_size": 32,
3703
- "bits": 4
3704
  },
3705
  "model.layers.45.mlp.shared_experts.down_proj": {
3706
  "group_size": 32,
3707
- "bits": 4
3708
  },
3709
  "lm_head": {
3710
  "group_size": 32,
 
1827
  },
1828
  "model.layers.45.self_attn.q_proj": {
1829
  "group_size": 32,
1830
+ "bits": 6
1831
  },
1832
  "model.layers.45.self_attn.k_proj": {
1833
  "group_size": 32,
1834
+ "bits": 6
1835
  },
1836
  "model.layers.45.self_attn.v_proj": {
1837
  "group_size": 32,
 
1839
  },
1840
  "model.layers.45.self_attn.o_proj": {
1841
  "group_size": 32,
1842
+ "bits": 6
1843
  },
1844
  "model.layers.45.mlp.switch_mlp.gate_proj": {
1845
  "group_size": 32,
 
1847
  },
1848
  "model.layers.45.mlp.switch_mlp.up_proj": {
1849
  "group_size": 32,
1850
+ "bits": 6
1851
  },
1852
  "model.layers.45.mlp.switch_mlp.down_proj": {
1853
  "group_size": 32,
1854
+ "bits": 6
1855
  },
1856
  "model.layers.45.mlp.shared_experts.gate_proj": {
1857
  "group_size": 32,
1858
+ "bits": 6
1859
  },
1860
  "model.layers.45.mlp.shared_experts.up_proj": {
1861
  "group_size": 32,
1862
+ "bits": 6
1863
  },
1864
  "model.layers.45.mlp.shared_experts.down_proj": {
1865
  "group_size": 32,
1866
+ "bits": 6
1867
  },
1868
  "lm_head": {
1869
  "group_size": 32,
 
3668
  },
3669
  "model.layers.45.self_attn.q_proj": {
3670
  "group_size": 32,
3671
+ "bits": 6
3672
  },
3673
  "model.layers.45.self_attn.k_proj": {
3674
  "group_size": 32,
3675
+ "bits": 6
3676
  },
3677
  "model.layers.45.self_attn.v_proj": {
3678
  "group_size": 32,
 
3680
  },
3681
  "model.layers.45.self_attn.o_proj": {
3682
  "group_size": 32,
3683
+ "bits": 6
3684
  },
3685
  "model.layers.45.mlp.switch_mlp.gate_proj": {
3686
  "group_size": 32,
 
3688
  },
3689
  "model.layers.45.mlp.switch_mlp.up_proj": {
3690
  "group_size": 32,
3691
+ "bits": 6
3692
  },
3693
  "model.layers.45.mlp.switch_mlp.down_proj": {
3694
  "group_size": 32,
3695
+ "bits": 6
3696
  },
3697
  "model.layers.45.mlp.shared_experts.gate_proj": {
3698
  "group_size": 32,
3699
+ "bits": 6
3700
  },
3701
  "model.layers.45.mlp.shared_experts.up_proj": {
3702
  "group_size": 32,
3703
+ "bits": 6
3704
  },
3705
  "model.layers.45.mlp.shared_experts.down_proj": {
3706
  "group_size": 32,
3707
+ "bits": 6
3708
  },
3709
  "lm_head": {
3710
  "group_size": 32,
model-00012-of-00012.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:24e1d9df387fc4677150a37d585027fd31dba1cc772a65ce11dd3a2ca75842ea
3
- size 2158177728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4d3e33f99555de00ab1172ee77d428681ac67516f9855840b79977c35d438c4
3
+ size 2465541568
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 59263583104,
4
  "total_parameters": 81932185824
5
  },
6
  "weight_map": {
 
1
  {
2
  "metadata": {
3
+ "total_size": 59570946944,
4
  "total_parameters": 81932185824
5
  },
6
  "weight_map": {