jiegcheng commited on
Commit
860c100
·
verified ·
1 Parent(s): 618dd6d

End of training

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: Qwen/Qwen3-1.7B
3
  library_name: transformers
4
  model_name: cm_adapter
5
  tags:
@@ -11,7 +11,7 @@ licence: license
11
 
12
  # Model Card for cm_adapter
13
 
14
- This model is a fine-tuned version of [Qwen/Qwen3-1.7B](https://huggingface.co/Qwen/Qwen3-1.7B).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
@@ -36,7 +36,7 @@ This model was trained with SFT.
36
 
37
  - TRL: 0.18.1
38
  - Transformers: 4.52.4
39
- - Pytorch: 2.5.1+cu121
40
  - Datasets: 3.6.0
41
  - Tokenizers: 0.21.1
42
 
 
1
  ---
2
+ base_model: Qwen/Qwen3-8B
3
  library_name: transformers
4
  model_name: cm_adapter
5
  tags:
 
11
 
12
  # Model Card for cm_adapter
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen3-8B](https://huggingface.co/Qwen/Qwen3-8B).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
36
 
37
  - TRL: 0.18.1
38
  - Transformers: 4.52.4
39
+ - Pytorch: 2.7.1
40
  - Datasets: 3.6.0
41
  - Tokenizers: 0.21.1
42
 
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "Qwen/Qwen3-1.7B",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -24,12 +24,12 @@
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "o_proj",
28
  "v_proj",
29
- "down_proj",
30
  "gate_proj",
 
31
  "up_proj",
32
- "q_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "trainable_token_indices": null,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen3-8B",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
+ "q_proj",
28
  "v_proj",
 
29
  "gate_proj",
30
+ "o_proj",
31
  "up_proj",
32
+ "down_proj"
33
  ],
34
  "task_type": "CAUSAL_LM",
35
  "trainable_token_indices": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7997a03bbcf90a0b63b6e7350bd1813021dc56ccb584bfd925926ad727f4c32
3
- size 64269832
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88ecfda7e503333ec0bc0046e813e37082b2a55f3b0f0633a328d7e74a925ea9
3
+ size 162849256
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:cbebc83a66c33eec1fc8453c5394ee6d3809a5c27acb90d51a71567fbef5feb6
3
- size 5624
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69dd12af6b063a4bd155ac751bad13d8da2c0ad3eacf7b1610ce4f993ab9e887
3
+ size 6033