bobchenyx commited on
Commit
f223ed8
·
verified ·
1 Parent(s): 307a739

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -3
README.md CHANGED
@@ -1,3 +1,41 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ quantized_by: bobchenyx
3
+ base_model:
4
+ - zai-org/GLM-4.6
5
+ base_model_relation: quantized
6
+ license: mit
7
+ tags:
8
+ - GLM
9
+ - GLM-4.6
10
+ - transformers
11
+ - GGUF
12
+ pipeline_tag: text-generation
13
+ ---
14
+
15
+ ## Llamacpp Quantizations of zai-org/GLM-4.6
16
+
17
+ Adopting **BF16** & **Imatrix** from [unsloth/GLM-4.6-GGUF](https://huggingface.co/unsloth/GLM-4.6-GGUF). (Huge fan of unsloth)
18
+
19
+
20
+ Personalized Replication of Low-Bit Mixed Precision Quant using `--tensor-type` option in [llama.cpp](https://github.com/ggml-org/llama.cpp)
21
+
22
+ ```
23
+ - IQ1_M : 83.63 GiB (2.01 BPW)
24
+ ```
25
+
26
+
27
+
28
+ ---
29
+ ## Download Guide
30
+
31
+ ```
32
+ # !pip install huggingface_hub hf_transfer
33
+ import os
34
+ os.environ["HF_HUB_ENABLE_HF_TRANSFER"] = "1"
35
+ from huggingface_hub import snapshot_download
36
+ snapshot_download(
37
+ repo_id = "bobchenyx/GLM-4.6-GGUF",
38
+ local_dir = "bobchenyx/GLM-4.6-GGUF",
39
+ allow_patterns = ["*IQ1_M*"], # Q2_K_L,Q4_K_M
40
+ )
41
+ ```