danielhanchen commited on
Commit
c5e26ab
·
verified ·
1 Parent(s): 1877031

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -103,3 +103,7 @@ Qwen3-235B-A22B-Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
103
  Qwen3-235B-A22B-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
104
  Qwen3-235B-A22B-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text
105
  Qwen3-235B-A22B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
103
  Qwen3-235B-A22B-Q4_0.gguf filter=lfs diff=lfs merge=lfs -text
104
  Qwen3-235B-A22B-Q4_1.gguf filter=lfs diff=lfs merge=lfs -text
105
  Qwen3-235B-A22B-Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
106
+ UD-Q5_K_XL/Qwen3-235B-A22B-UD-Q5_K_XL-00004-of-00004.gguf filter=lfs diff=lfs merge=lfs -text
107
+ UD-Q5_K_XL/Qwen3-235B-A22B-UD-Q5_K_XL-00001-of-00004.gguf filter=lfs diff=lfs merge=lfs -text
108
+ UD-Q5_K_XL/Qwen3-235B-A22B-UD-Q5_K_XL-00003-of-00004.gguf filter=lfs diff=lfs merge=lfs -text
109
+ UD-Q5_K_XL/Qwen3-235B-A22B-UD-Q5_K_XL-00002-of-00004.gguf filter=lfs diff=lfs merge=lfs -text
Q4_K_M/Qwen3-235B-A22B-Q4_K_M-00001-of-00003.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:41ca28701b71704d0ecf89f38c4b643c0b7a322bfc4d893d079ab08bc05b5820
3
- size 49944699552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:540a34b4540ee2586f3565d11f5758bbef405ae915eac38ffc4ac7ba1360022c
3
+ size 49944699712
UD-Q2_K_XL/Qwen3-235B-A22B-UD-Q2_K_XL-00001-of-00002.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:43ea4d68c138800c68903644d71fdd751528a60e620c87f7c56609600a5105e6
3
- size 49841583232
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a71fe713905dd8ae4eed73d08ad678efaa3a3d39eecc0211ec8b49551ef60315
3
+ size 49841583360
UD-Q3_K_XL/Qwen3-235B-A22B-UD-Q3_K_XL-00001-of-00003.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:749d995b33573187f86c2e1dafdcf4d18940f1aa6bc88128e081e1f5099008cb
3
- size 49966873792
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c84ecae31bfab63449cb33c6ac514fcc0887e0e3c522da7956e73fee635fe0b
3
+ size 49966873952
UD-Q4_K_XL/Qwen3-235B-A22B-UD-Q4_K_XL-00001-of-00003.gguf CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca6d7c07142c82cf17d0e53ffebb5ad8dab75a86a202690879cc76da596a811f
3
- size 49875808608
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:633a1a7b6c3c4e2c53f9f3cfeedb4527a79aeda9d88e767e6a71a9a0dfffd06f
3
+ size 49875808736
UD-Q5_K_XL/Qwen3-235B-A22B-UD-Q5_K_XL-00001-of-00004.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2adedf41f5224917f2e71435a41e9b8f27873ea0dd2aa719c0f5143bcc99228a
3
+ size 49835132704
UD-Q5_K_XL/Qwen3-235B-A22B-UD-Q5_K_XL-00002-of-00004.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ba5811ce74a055b8d45b669a32e8fa0075a92258f82957c4aad20c8492d94b1
3
+ size 49569711008
UD-Q5_K_XL/Qwen3-235B-A22B-UD-Q5_K_XL-00003-of-00004.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:158472f3f092219508d2ce1d9f7a5ca7c45550841aa50af03a7d41e3d6ae48c8
3
+ size 49678779392
UD-Q5_K_XL/Qwen3-235B-A22B-UD-Q5_K_XL-00004-of-00004.gguf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7916d58e3392f1b017c14d05eaaf997ee7b19130c5ebe413c9fc01ce89706289
3
+ size 17802198880
config.json CHANGED
@@ -24,6 +24,7 @@
24
  "num_hidden_layers": 94,
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
 
27
  "rms_norm_eps": 1e-06,
28
  "rope_scaling": null,
29
  "rope_theta": 1000000.0,
@@ -31,8 +32,9 @@
31
  "sliding_window": null,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
34
- "transformers_version": "4.51.0",
 
35
  "use_cache": true,
36
  "use_sliding_window": false,
37
  "vocab_size": 151936
38
- }
 
24
  "num_hidden_layers": 94,
25
  "num_key_value_heads": 4,
26
  "output_router_logits": false,
27
+ "pad_token_id": 151654,
28
  "rms_norm_eps": 1e-06,
29
  "rope_scaling": null,
30
  "rope_theta": 1000000.0,
 
32
  "sliding_window": null,
33
  "tie_word_embeddings": false,
34
  "torch_dtype": "bfloat16",
35
+ "transformers_version": "4.51.3",
36
+ "unsloth_fixed": true,
37
  "use_cache": true,
38
  "use_sliding_window": false,
39
  "vocab_size": 151936
40
+ }