Update README.md
Browse files
README.md
CHANGED
@@ -46,6 +46,19 @@ For instance, serve the model as follows:
|
|
46 |
vllm serve RedHatAI/Magistral-Small-2506-FP8 --tokenizer-mode mistral --config-format mistral --load-format mistral --tool-call-parser mistral --enable-auto-tool-choice
|
47 |
```
|
48 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
49 |
# Original Model Card
|
50 |
|
51 |
Building upon Mistral Small 3.1 (2503), **with added reasoning capabilities**, undergoing SFT from Magistral Medium traces and RL on top, it's a small, efficient reasoning model with 24B parameters.
|
|
|
46 |
vllm serve RedHatAI/Magistral-Small-2506-FP8 --tokenizer-mode mistral --config-format mistral --load-format mistral --tool-call-parser mistral --enable-auto-tool-choice
|
47 |
```
|
48 |
|
49 |
+
## Evaluation
|
50 |
+
|
51 |
+
GSM8k:
|
52 |
+
|
53 |
+
```
|
54 |
+
lm_eval --model local-completions --model_args model=RedHatAI/Magistral-Small-2506-FP8,base_url=http://0.0.0.0:9000/v1/completions,num_concurrent=500,tokenized_requests=False --tasks gsm8k --num_fewshot 5
|
55 |
+
local-completions (model=RedHatAI/Magistral-Small-2506-FP8,base_url=http://0.0.0.0:9000/v1/completions,num_concurrent=500,tokenized_requests=False), gen_kwargs: (None), limit: None, num_fewshot: 5, batch_size: 1
|
56 |
+
|Tasks|Version| Filter |n-shot| Metric | |Value | |Stderr|
|
57 |
+
|-----|------:|----------------|-----:|-----------|---|-----:|---|-----:|
|
58 |
+
|gsm8k| 3|flexible-extract| 5|exact_match|↑ |0.8923|± |0.0085|
|
59 |
+
| | |strict-match | 5|exact_match|↑ |0.8886|± |0.0087|
|
60 |
+
```
|
61 |
+
|
62 |
# Original Model Card
|
63 |
|
64 |
Building upon Mistral Small 3.1 (2503), **with added reasoning capabilities**, undergoing SFT from Magistral Medium traces and RL on top, it's a small, efficient reasoning model with 24B parameters.
|