Felladrin commited on
Commit
b2167c7
·
verified ·
1 Parent(s): a871198

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,110 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers.js
3
+ license: apache-2.0
4
+ datasets:
5
+ - HuggingFaceH4/ultrachat_200k
6
+ language:
7
+ - en
8
+ base_model:
9
+ - Felladrin/Minueza-2-96M-Instruct-Variant-10
10
+ tags:
11
+ - llama-factory
12
+ pipeline_tag: text-generation
13
+ ---
14
+
15
+
16
+
17
+ # Minueza-2-96M-Instruct-Variant-10 (ONNX)
18
+
19
+
20
+ This is an ONNX version of [Felladrin/Minueza-2-96M-Instruct-Variant-10](https://huggingface.co/Felladrin/Minueza-2-96M-Instruct-Variant-10). It was automatically converted and uploaded using [this Hugging Face Space](https://huggingface.co/spaces/onnx-community/convert-to-onnx).
21
+
22
+
23
+ ## Usage with Transformers.js
24
+
25
+
26
+ See the pipeline documentation for `text-generation`: https://huggingface.co/docs/transformers.js/api/pipelines#module_pipelines.TextGenerationPipeline
27
+
28
+
29
+ ---
30
+
31
+
32
+ # Minueza-2-96M-Instruct (Variant 10)
33
+
34
+ This model is a fine-tuned version of [Felladrin/Minueza-2-96M](https://huggingface.co/Felladrin/Minueza-2-96M) on the English [HuggingFaceH4/ultrachat_200k](https://huggingface.co/datasets/HuggingFaceH4/ultrachat_200k) dataset.
35
+
36
+ ## Usage
37
+
38
+ ```sh
39
+ pip install transformers==4.51.1 torch==2.6.0
40
+ ```
41
+
42
+ ```python
43
+ from transformers import pipeline, TextStreamer
44
+ import torch
45
+
46
+ generate_text = pipeline(
47
+ "text-generation",
48
+ model="Felladrin/Minueza-2-96M-Instruct-Variant-10",
49
+ device=torch.device("cuda" if torch.cuda.is_available() else "cpu"),
50
+ )
51
+
52
+ messages = [
53
+ {
54
+ "role": "system",
55
+ "content": "You are a career counselor. The user will provide you with an individual looking for guidance in their professional life, and your task is to assist them in determining what careers they are most suited for based on their skills, interests, and experience. You should also conduct research into the various options available, explain the job market trends in different industries, and advice on which qualifications would be beneficial for pursuing particular fields.",
56
+ },
57
+ {
58
+ "role": "user",
59
+ "content": "Hi!",
60
+ },
61
+ {
62
+ "role": "assistant",
63
+ "content": "Hello! How can I help you?",
64
+ },
65
+ {
66
+ "role": "user",
67
+ "content": "I am interested in developing a career in software engineering. Do you have any suggestions?",
68
+ },
69
+ ]
70
+
71
+ generate_text(
72
+ generate_text.tokenizer.apply_chat_template(
73
+ messages, tokenize=False, add_generation_prompt=True
74
+ ),
75
+ streamer=TextStreamer(generate_text.tokenizer, skip_special_tokens=True),
76
+ max_new_tokens=512,
77
+ do_sample=True,
78
+ temperature=0.7,
79
+ top_p=0.9,
80
+ top_k=0,
81
+ min_p=0.1,
82
+ repetition_penalty=1.17,
83
+ )
84
+ ```
85
+
86
+ ## Training hyperparameters
87
+
88
+ The following hyperparameters were used during training:
89
+
90
+ - learning_rate: 5.8e-05
91
+ - train_batch_size: 4
92
+ - eval_batch_size: 4
93
+ - seed: 42
94
+ - gradient_accumulation_steps: 32
95
+ - total_train_batch_size: 128
96
+ - optimizer: Use adamw_torch with betas=(0.9,0.95) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
97
+ - lr_scheduler_type: cosine
98
+ - lr_scheduler_warmup_ratio: 0.1
99
+ - num_epochs: 2
100
+
101
+ ## Framework versions
102
+
103
+ - Transformers 4.51.1
104
+ - Pytorch 2.6.0+cu124
105
+ - Datasets 3.5.0
106
+ - Tokenizers 0.21.0
107
+
108
+ ## License
109
+
110
+ This model is licensed under the Apache License 2.0.
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_attn_implementation_autoset": true,
3
+ "_name_or_path": "Felladrin/Minueza-2-96M-Instruct-Variant-10",
4
+ "architectures": [
5
+ "LlamaForCausalLM"
6
+ ],
7
+ "attention_bias": false,
8
+ "attention_dropout": 0.1,
9
+ "bos_token_id": 1,
10
+ "eos_token_id": 2,
11
+ "head_dim": 56,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 672,
14
+ "initializer_range": 0.02,
15
+ "intermediate_size": 2688,
16
+ "max_position_embeddings": 4096,
17
+ "mlp_bias": false,
18
+ "model_type": "llama",
19
+ "num_attention_heads": 12,
20
+ "num_hidden_layers": 8,
21
+ "num_key_value_heads": 4,
22
+ "pretraining_tp": 1,
23
+ "rms_norm_eps": 1e-06,
24
+ "rope_scaling": null,
25
+ "rope_theta": 500000.0,
26
+ "tie_word_embeddings": false,
27
+ "torch_dtype": "float32",
28
+ "transformers_version": "4.49.0",
29
+ "use_cache": false,
30
+ "vocab_size": 32000
31
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "eos_token_id": 2,
5
+ "transformers_version": "4.49.0"
6
+ }
onnx/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e2dbcc40e8eb5f9f475a171aff85b679b07f6a7e68fae70963238812a5e00c8
3
+ size 384240317
onnx/model_bnb4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee324dee78b4b4ce1f7307e71037716f31ae3a0b8759c6f4e5db18bb6570ead6
3
+ size 128190351
onnx/model_fp16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:28c1d9a241409e81a2d3784552c2ade38d275cb3a640815a4c3afebbfc631e65
3
+ size 192230262
onnx/model_int8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92cf845c1d227ffa1e5cc3621ac5a520f78a4f1346cf26f0b4d2cb2f85752c85
3
+ size 96320555
onnx/model_q4.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:be7e3038b88b08a9a5aca6add207ee30b7da6bbf34012a7b7b2b6b20a71feff9
3
+ size 132845545
onnx/model_q4f16.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fcf34f106c53f49a2adfb5f5e6af98e876667f359d81088c1d4c146a5cda9d3
3
+ size 85159553
onnx/model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92cf845c1d227ffa1e5cc3621ac5a520f78a4f1346cf26f0b4d2cb2f85752c85
3
+ size 96320555
onnx/model_uint8.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e3a414d1504c8c86368b496f144c7b9bd0a67c55350fc97b58c1219f2260767
3
+ size 96320586
quantize_config.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "modes": [
3
+ "fp16",
4
+ "q8",
5
+ "int8",
6
+ "uint8",
7
+ "q4",
8
+ "q4f16",
9
+ "bnb4"
10
+ ],
11
+ "per_channel": false,
12
+ "reduce_range": false,
13
+ "block_size": null,
14
+ "is_symmetric": true,
15
+ "accuracy_level": null,
16
+ "quant_type": 1,
17
+ "op_block_list": null
18
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<|im_start|>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "<|im_end|>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<|im_end|>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,40 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<|im_start|>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<|im_end|>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "bos_token": "<|im_start|>",
29
+ "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a highly knowledgeable and friendly assistant. Your goal is to understand and respond to user inquiries with clarity. Your interactions are always respectful, helpful, and focused on delivering the most accurate information to the user.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
30
+ "clean_up_tokenization_spaces": false,
31
+ "eos_token": "<|im_end|>",
32
+ "extra_special_tokens": {},
33
+ "model_max_length": 4096,
34
+ "pad_token": "<|im_end|>",
35
+ "padding_side": "right",
36
+ "split_special_tokens": false,
37
+ "tokenizer_class": "PreTrainedTokenizer",
38
+ "truncation_side": "right",
39
+ "unk_token": "<unk>"
40
+ }