Gustavo de Rosa
commited on
Commit
·
57faa53
1
Parent(s):
34d6e8c
chore(root): Adds top_k information even if 50 is already the default.
Browse files- README.md +2 -1
- generation_config.json +1 -0
README.md
CHANGED
|
@@ -56,7 +56,7 @@ library_name: transformers
|
|
| 56 |
## Usage
|
| 57 |
|
| 58 |
> [!IMPORTANT]
|
| 59 |
-
> To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
|
| 60 |
|
| 61 |
### Input Formats
|
| 62 |
|
|
@@ -88,6 +88,7 @@ outputs = model.generate(
|
|
| 88 |
inputs.to(model.device),
|
| 89 |
max_new_tokens=4096,
|
| 90 |
temperature=0.8,
|
|
|
|
| 91 |
top_p=0.95,
|
| 92 |
do_sample=True,
|
| 93 |
)
|
|
|
|
| 56 |
## Usage
|
| 57 |
|
| 58 |
> [!IMPORTANT]
|
| 59 |
+
> To fully take advantage of the model's capabilities, inference must use `temperature=0.8`, `top_k=50`, `top_p=0.95`, and `do_sample=True`. For more complex queries, set `max_new_tokens=32768` to allow for longer chain-of-thought (CoT).
|
| 60 |
|
| 61 |
### Input Formats
|
| 62 |
|
|
|
|
| 88 |
inputs.to(model.device),
|
| 89 |
max_new_tokens=4096,
|
| 90 |
temperature=0.8,
|
| 91 |
+
top_k=50,
|
| 92 |
top_p=0.95,
|
| 93 |
do_sample=True,
|
| 94 |
)
|
generation_config.json
CHANGED
|
@@ -5,6 +5,7 @@
|
|
| 5 |
"eos_token_id": 100265,
|
| 6 |
"pad_token_id": 100349,
|
| 7 |
"temperature": 0.8,
|
|
|
|
| 8 |
"top_p": 0.95,
|
| 9 |
"transformers_version": "4.51.1"
|
| 10 |
}
|
|
|
|
| 5 |
"eos_token_id": 100265,
|
| 6 |
"pad_token_id": 100349,
|
| 7 |
"temperature": 0.8,
|
| 8 |
+
"top_k": 50,
|
| 9 |
"top_p": 0.95,
|
| 10 |
"transformers_version": "4.51.1"
|
| 11 |
}
|