Dataset Viewer
eval_name
stringlengths 12
111
| Precision
stringclasses 3
values | Type
stringclasses 6
values | T
stringclasses 6
values | Weight type
stringclasses 2
values | Architecture
stringclasses 52
values | Model
stringlengths 355
689
| fullname
stringlengths 4
102
| Model sha
stringlengths 0
40
| Average ⬆️
float64 1.03
52
| Hub License
stringclasses 26
values | Hub ❤️
int64 0
5.9k
| #Params (B)
int64 -1
140
| Available on the hub
bool 2
classes | MoE
bool 2
classes | Flagged
bool 2
classes | Chat Template
bool 2
classes | CO₂ cost (kg)
float64 0.03
107
| IFEval Raw
float64 0
0.9
| IFEval
float64 0
90
| BBH Raw
float64 0.27
0.75
| BBH
float64 0.81
63.5
| MATH Lvl 5 Raw
float64 0
0.51
| MATH Lvl 5
float64 0
50.7
| GPQA Raw
float64 0.22
0.44
| GPQA
float64 0
24.9
| MUSR Raw
float64 0.29
0.6
| MUSR
float64 0
38.5
| MMLU-PRO Raw
float64 0.1
0.73
| MMLU-PRO
float64 0
70
| Merged
bool 2
classes | Official Providers
bool 2
classes | Upload To Hub Date
stringclasses 424
values | Submission Date
stringclasses 169
values | Generation
int64 0
10
| Base Model
stringlengths 4
102
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0-hero_Matter-0.2-7B-DPO_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/0-hero/Matter-0.2-7B-DPO" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">0-hero/Matter-0.2-7B-DPO</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/0-hero__Matter-0.2-7B-DPO-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
0-hero/Matter-0.2-7B-DPO
|
26a66f0d862e2024ce4ad0a09c37052ac36e8af6
| 8.805656 |
apache-2.0
| 3 | 7 | true | false | false | true | 0.609587 | 0.330279 | 33.027921 | 0.359625 | 10.055525 | 0.008308 | 0.830816 | 0.259228 | 1.230425 | 0.381375 | 5.871875 | 0.116356 | 1.817376 | false | false |
2024-04-13
|
2024-08-05
| 0 |
0-hero/Matter-0.2-7B-DPO
|
01-ai_Yi-1.5-34B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-34B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-34B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-34B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-34B
|
4b486f81c935a2dadde84c6baa1e1370d40a098f
| 25.621318 |
apache-2.0
| 47 | 34 | true | false | false | false | 11.351699 | 0.284117 | 28.411725 | 0.597639 | 42.749363 | 0.151813 | 15.181269 | 0.365772 | 15.436242 | 0.423604 | 11.217188 | 0.466589 | 40.732122 | false | true |
2024-05-11
|
2024-06-12
| 0 |
01-ai/Yi-1.5-34B
|
01-ai_Yi-1.5-34B-32K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-34B-32K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-34B-32K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-34B-32K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-34B-32K
|
2c03a29761e4174f20347a60fbe229be4383d48b
| 26.67756 |
apache-2.0
| 36 | 34 | true | false | false | false | 11.577314 | 0.311869 | 31.186917 | 0.601569 | 43.381847 | 0.151057 | 15.10574 | 0.363255 | 15.100671 | 0.439823 | 14.077865 | 0.470911 | 41.212323 | false | true |
2024-05-15
|
2024-06-12
| 0 |
01-ai/Yi-1.5-34B-32K
|
01-ai_Yi-1.5-34B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-34B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-34B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-34B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-34B-Chat
|
f3128b2d02d82989daae566c0a7eadc621ca3254
| 32.892233 |
apache-2.0
| 257 | 34 | true | false | false | true | 11.211922 | 0.606676 | 60.667584 | 0.608375 | 44.262826 | 0.249245 | 24.924471 | 0.364933 | 15.324385 | 0.428198 | 13.058073 | 0.452045 | 39.116061 | false | true |
2024-05-10
|
2024-06-12
| 0 |
01-ai/Yi-1.5-34B-Chat
|
01-ai_Yi-1.5-34B-Chat-16K_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-34B-Chat-16K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-34B-Chat-16K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-34B-Chat-16K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-34B-Chat-16K
|
ff74452e11f0f749ab872dc19b1dd3813c25c4d8
| 29.239909 |
apache-2.0
| 27 | 34 | true | false | false | true | 3.387011 | 0.45645 | 45.645 | 0.610022 | 44.536157 | 0.203927 | 20.392749 | 0.338087 | 11.744966 | 0.43976 | 13.736719 | 0.454455 | 39.383865 | false | true |
2024-05-15
|
2024-07-15
| 0 |
01-ai/Yi-1.5-34B-Chat-16K
|
01-ai_Yi-1.5-6B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-6B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-6B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-6B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-6B
|
cab51fce425b4c1fb19fccfdd96bd5d0908c1657
| 16.695346 |
apache-2.0
| 30 | 6 | true | false | false | false | 1.184757 | 0.26166 | 26.166017 | 0.449258 | 22.027905 | 0.063444 | 6.344411 | 0.313758 | 8.501119 | 0.437406 | 13.309115 | 0.314412 | 23.823508 | false | true |
2024-05-11
|
2024-08-10
| 0 |
01-ai/Yi-1.5-6B
|
01-ai_Yi-1.5-6B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-6B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-6B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-6B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-6B-Chat
|
3f64d3f159c6ad8494227bb77e2a7baef8cd808b
| 20.983906 |
apache-2.0
| 41 | 6 | true | false | false | true | 0.96884 | 0.514527 | 51.452701 | 0.457131 | 23.678723 | 0.054381 | 5.438066 | 0.302013 | 6.935123 | 0.439177 | 14.030469 | 0.319315 | 24.368351 | false | true |
2024-05-11
|
2024-10-22
| 0 |
01-ai/Yi-1.5-6B-Chat
|
01-ai_Yi-1.5-9B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-9B
|
8cfde9604384c50137bee480b8cef8a08e5ae81d
| 22.141313 |
apache-2.0
| 47 | 8 | true | false | false | false | 0.734446 | 0.293584 | 29.358436 | 0.514294 | 30.500717 | 0.113293 | 11.329305 | 0.379195 | 17.225951 | 0.432781 | 12.03099 | 0.391622 | 32.402482 | false | true |
2024-05-11
|
2024-06-12
| 0 |
01-ai/Yi-1.5-9B
|
01-ai_Yi-1.5-9B-32K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-9B-32K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-9B-32K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-9B-32K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-9B-32K
|
116561dfae63af90f9d163b43077629e0e916bb1
| 19.78461 |
apache-2.0
| 18 | 8 | true | false | false | false | 0.784037 | 0.230311 | 23.031113 | 0.496332 | 28.937012 | 0.106495 | 10.649547 | 0.35906 | 14.541387 | 0.418615 | 10.826823 | 0.376496 | 30.721779 | false | true |
2024-05-15
|
2024-06-12
| 0 |
01-ai/Yi-1.5-9B-32K
|
01-ai_Yi-1.5-9B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-9B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-9B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-9B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-9B-Chat
|
bc87d8557c98dc1e5fdef6ec23ed31088c4d3f35
| 27.894417 |
apache-2.0
| 135 | 8 | true | false | false | true | 0.726772 | 0.604553 | 60.455259 | 0.555906 | 36.952931 | 0.127644 | 12.76435 | 0.334732 | 11.297539 | 0.425906 | 12.838281 | 0.397523 | 33.058141 | false | true |
2024-05-10
|
2024-06-12
| 0 |
01-ai/Yi-1.5-9B-Chat
|
01-ai_Yi-1.5-9B-Chat-16K_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-1.5-9B-Chat-16K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-1.5-9B-Chat-16K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-1.5-9B-Chat-16K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-1.5-9B-Chat-16K
|
2b397e5f0fab87984efa66856c5c4ed4bbe68b50
| 23.035282 |
apache-2.0
| 34 | 8 | true | false | false | true | 0.792373 | 0.421404 | 42.14041 | 0.515338 | 31.497609 | 0.134441 | 13.444109 | 0.308725 | 7.829978 | 0.409906 | 10.038281 | 0.399352 | 33.261303 | false | true |
2024-05-15
|
2024-06-12
| 0 |
01-ai/Yi-1.5-9B-Chat-16K
|
01-ai_Yi-34B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-34B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-34B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-34B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-34B
|
e1e7da8c75cfd5c44522228599fd4d2990cedd1c
| 22.385715 |
apache-2.0
| 1,287 | 34 | true | false | false | false | 12.828742 | 0.304575 | 30.457519 | 0.54571 | 35.542431 | 0.052115 | 5.21148 | 0.366611 | 15.548098 | 0.411854 | 9.648438 | 0.441157 | 37.906324 | false | true |
2023-11-01
|
2024-06-12
| 0 |
01-ai/Yi-34B
|
01-ai_Yi-34B-200K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-34B-200K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-34B-200K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-34B-200K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-34B-200K
|
8ac1a1ebe011df28b78ccd08012aeb2222443c77
| 19.887594 |
apache-2.0
| 316 | 34 | true | false | false | false | 12.751928 | 0.154249 | 15.424851 | 0.544182 | 36.02211 | 0.049849 | 4.984894 | 0.356544 | 14.205817 | 0.381719 | 9.414844 | 0.453457 | 39.27305 | false | true |
2023-11-06
|
2024-06-12
| 0 |
01-ai/Yi-34B-200K
|
01-ai_Yi-34B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-34B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-34B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-34B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-34B-Chat
|
2e528b6a80fb064a0a746c5ca43114b135e30464
| 23.962312 |
apache-2.0
| 345 | 34 | true | false | false | true | 12.562848 | 0.469889 | 46.988878 | 0.556087 | 37.623988 | 0.046828 | 4.682779 | 0.338087 | 11.744966 | 0.397844 | 8.363802 | 0.409325 | 34.369459 | false | true |
2023-11-22
|
2024-06-12
| 0 |
01-ai/Yi-34B-Chat
|
01-ai_Yi-6B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-6B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-6B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-6B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-6B
|
7f7fb7662fd8ec09029364f408053c954986c8e5
| 13.611617 |
apache-2.0
| 371 | 6 | true | false | false | false | 0.549275 | 0.289338 | 28.933785 | 0.430923 | 19.408505 | 0.015861 | 1.586103 | 0.269295 | 2.572707 | 0.393687 | 7.044271 | 0.299119 | 22.124335 | false | true |
2023-11-01
|
2024-06-12
| 0 |
01-ai/Yi-6B
|
01-ai_Yi-6B-200K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-6B-200K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-6B-200K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-6B-200K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-6B-200K
|
4a74338e778a599f313e9fa8f5bc08c717604420
| 11.933158 |
apache-2.0
| 173 | 6 | true | false | false | false | 0.563212 | 0.084331 | 8.433069 | 0.428929 | 20.14802 | 0.01435 | 1.435045 | 0.281879 | 4.250559 | 0.45874 | 16.842448 | 0.284408 | 20.489805 | false | true |
2023-11-06
|
2024-06-12
| 0 |
01-ai/Yi-6B-200K
|
01-ai_Yi-6B-Chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-6B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-6B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-6B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-6B-Chat
|
01f7fabb6cfb26efeb764da4a0a19cad2c754232
| 14.004357 |
apache-2.0
| 63 | 6 | true | false | false | true | 0.555333 | 0.339521 | 33.952136 | 0.41326 | 17.000167 | 0.006798 | 0.679758 | 0.294463 | 5.928412 | 0.368792 | 3.565625 | 0.3061 | 22.900044 | false | true |
2023-11-22
|
2024-06-12
| 0 |
01-ai/Yi-6B-Chat
|
01-ai_Yi-9B_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-9B
|
b4a466d95091696285409f1dcca3028543cb39da
| 17.774103 |
apache-2.0
| 186 | 8 | true | false | false | false | 0.765332 | 0.270878 | 27.087794 | 0.493961 | 27.626956 | 0.053625 | 5.362538 | 0.317953 | 9.060403 | 0.405406 | 8.909115 | 0.35738 | 28.597813 | false | true |
2024-03-01
|
2024-06-12
| 0 |
01-ai/Yi-9B
|
01-ai_Yi-9B-200K_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-9B-200K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-9B-200K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-9B-200K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-9B-200K
|
8c93accd5589dbb74ee938e103613508c4a9b88d
| 17.74214 |
apache-2.0
| 75 | 8 | true | false | false | false | 0.774491 | 0.232709 | 23.270921 | 0.47933 | 26.492495 | 0.067221 | 6.722054 | 0.315436 | 8.724832 | 0.429406 | 12.109115 | 0.362201 | 29.133422 | false | true |
2024-03-15
|
2024-06-12
| 0 |
01-ai/Yi-9B-200K
|
01-ai_Yi-Coder-9B-Chat_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/01-ai/Yi-Coder-9B-Chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">01-ai/Yi-Coder-9B-Chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/01-ai__Yi-Coder-9B-Chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
01-ai/Yi-Coder-9B-Chat
|
356a1f8d4e4a606d0b879e54191ca809918576b8
| 16.872696 |
apache-2.0
| 191 | 8 | true | false | false | true | 0.909766 | 0.481704 | 48.17041 | 0.48142 | 25.943153 | 0.033233 | 3.323263 | 0.247483 | 0 | 0.399177 | 7.963802 | 0.24252 | 15.83555 | false | true |
2024-08-21
|
2024-09-12
| 1 |
01-ai/Yi-Coder-9B
|
152334H_miqu-1-70b-sf_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/152334H/miqu-1-70b-sf" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">152334H/miqu-1-70b-sf</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/152334H__miqu-1-70b-sf-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
152334H/miqu-1-70b-sf
|
1dca4cce36f01f2104ee2e6b97bac6ff7bb300c1
| 29.059643 | 219 | 68 | false | false | false | false | 6.098986 | 0.518174 | 51.8174 | 0.610236 | 43.807147 | 0.122356 | 12.23565 | 0.350671 | 13.422819 | 0.458208 | 17.209375 | 0.422789 | 35.86547 | false | false |
2024-01-30
|
2024-06-26
| 0 |
152334H/miqu-1-70b-sf
|
|
1TuanPham_T-VisStar-7B-v0.1_float16
|
float16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/1TuanPham/T-VisStar-7B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">1TuanPham/T-VisStar-7B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/1TuanPham__T-VisStar-7B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
1TuanPham/T-VisStar-7B-v0.1
|
b111b59971c14b46c888b96723ff7f3c7b6fd92f
| 19.044104 |
apache-2.0
| 2 | 7 | true | false | false | true | 1.269513 | 0.360704 | 36.070404 | 0.50522 | 30.243834 | 0.05136 | 5.135952 | 0.285235 | 4.697987 | 0.4375 | 13.554167 | 0.321061 | 24.562278 | true | false |
2024-09-19
|
2024-09-22
| 0 |
1TuanPham/T-VisStar-7B-v0.1
|
1TuanPham_T-VisStar-v0.1_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/1TuanPham/T-VisStar-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">1TuanPham/T-VisStar-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/1TuanPham__T-VisStar-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
1TuanPham/T-VisStar-v0.1
|
c9779bd9630a533f7e42fd8effcca69623d48c9c
| 19.044104 |
apache-2.0
| 2 | 7 | true | false | false | true | 0.624384 | 0.360704 | 36.070404 | 0.50522 | 30.243834 | 0.05136 | 5.135952 | 0.285235 | 4.697987 | 0.4375 | 13.554167 | 0.321061 | 24.562278 | true | false |
2024-09-19
|
2024-09-20
| 0 |
1TuanPham/T-VisStar-v0.1
|
3rd-Degree-Burn_L-3.1-Science-Writer-8B_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/3rd-Degree-Burn/L-3.1-Science-Writer-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">3rd-Degree-Burn/L-3.1-Science-Writer-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/3rd-Degree-Burn__L-3.1-Science-Writer-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
3rd-Degree-Burn/L-3.1-Science-Writer-8B
|
d9bb11fb02f8eca3aec408912278e513377115da
| 21.07862 | 0 | 8 | false | false | false | false | 0.709678 | 0.42625 | 42.625013 | 0.504131 | 29.199301 | 0.102719 | 10.271903 | 0.274329 | 3.243848 | 0.395948 | 11.69349 | 0.364943 | 29.438165 | false | false |
2024-11-19
| 0 |
Removed
|
||
3rd-Degree-Burn_Llama-3.1-8B-Squareroot_float16
|
float16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/3rd-Degree-Burn/Llama-3.1-8B-Squareroot" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">3rd-Degree-Burn/Llama-3.1-8B-Squareroot</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/3rd-Degree-Burn__Llama-3.1-8B-Squareroot-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
3rd-Degree-Burn/Llama-3.1-8B-Squareroot
|
2bec01c2c5d53276eac2222c80190eb44ab2e6af
| 10.581747 |
apache-2.0
| 1 | 8 | true | false | false | true | 0.98705 | 0.221344 | 22.134381 | 0.346094 | 8.618064 | 0.227341 | 22.734139 | 0.256711 | 0.894855 | 0.308917 | 0.78125 | 0.17495 | 8.327793 | true | false |
2024-10-10
|
2024-10-10
| 1 |
3rd-Degree-Burn/Llama-3.1-8B-Squareroot (Merge)
|
3rd-Degree-Burn_Llama-3.1-8B-Squareroot-v1_float16
|
float16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/3rd-Degree-Burn/Llama-3.1-8B-Squareroot-v1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">3rd-Degree-Burn/Llama-3.1-8B-Squareroot-v1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/3rd-Degree-Burn__Llama-3.1-8B-Squareroot-v1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
3rd-Degree-Burn/Llama-3.1-8B-Squareroot-v1
|
09339d9c3b118ae3c6e7beab8b84347471990988
| 7.597362 | 0 | 8 | false | false | false | true | 0.772749 | 0.289238 | 28.923811 | 0.334277 | 6.515145 | 0.061934 | 6.193353 | 0.255872 | 0.782998 | 0.334063 | 1.757812 | 0.112699 | 1.411052 | false | false |
2024-11-10
| 0 |
Removed
|
||
3rd-Degree-Burn_Llama-Squared-8B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/3rd-Degree-Burn/Llama-Squared-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">3rd-Degree-Burn/Llama-Squared-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/3rd-Degree-Burn__Llama-Squared-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
3rd-Degree-Burn/Llama-Squared-8B
|
f30737e92b3a3fa0ef2a3f3ade487cc94ad34400
| 12.233544 | 0 | 8 | false | false | false | true | 1.011112 | 0.275524 | 27.55245 | 0.443103 | 21.277103 | 0.045317 | 4.531722 | 0.271812 | 2.908277 | 0.308948 | 1.951823 | 0.236619 | 15.179891 | false | false |
2024-10-08
| 0 |
Removed
|
||
4season_final_model_test_v2_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/4season/final_model_test_v2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">4season/final_model_test_v2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/4season__final_model_test_v2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
4season/final_model_test_v2
|
cf690c35d9cf0b0b6bf034fa16dbf88c56fe861c
| 21.91554 |
apache-2.0
| 0 | 21 | true | false | false | false | 1.081038 | 0.319113 | 31.911329 | 0.634205 | 47.41067 | 0.013595 | 1.359517 | 0.327181 | 10.290828 | 0.431448 | 12.43099 | 0.352809 | 28.089908 | false | false |
2024-05-20
|
2024-06-27
| 0 |
4season/final_model_test_v2
|
AALF_FuseChat-Llama-3.1-8B-Instruct-preview_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AALF/FuseChat-Llama-3.1-8B-Instruct-preview" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AALF/FuseChat-Llama-3.1-8B-Instruct-preview</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AALF__FuseChat-Llama-3.1-8B-Instruct-preview-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AALF/FuseChat-Llama-3.1-8B-Instruct-preview
|
f740497979293c90fa1cfaa7c446016e107cc2c1
| 25.610368 | 0 | 8 | false | false | false | true | 0.688619 | 0.718958 | 71.895792 | 0.511989 | 30.848065 | 0.070242 | 7.024169 | 0.305369 | 7.38255 | 0.382 | 6.15 | 0.373255 | 30.361628 | false | false |
2024-11-20
| 0 |
Removed
|
||
AALF_FuseChat-Llama-3.1-8B-SFT-preview_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AALF/FuseChat-Llama-3.1-8B-SFT-preview" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AALF/FuseChat-Llama-3.1-8B-SFT-preview</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AALF__FuseChat-Llama-3.1-8B-SFT-preview-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AALF/FuseChat-Llama-3.1-8B-SFT-preview
|
601f2b8c448acc5686656d3979ed732ce050b827
| 27.374839 | 0 | 8 | false | false | false | true | 0.684308 | 0.72805 | 72.805046 | 0.52403 | 32.536782 | 0.114048 | 11.404834 | 0.30453 | 7.270694 | 0.402 | 9.75 | 0.374335 | 30.481678 | false | false |
2024-11-21
| 0 |
Removed
|
||
AALF_gemma-2-27b-it-SimPO-37K_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AALF/gemma-2-27b-it-SimPO-37K" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AALF/gemma-2-27b-it-SimPO-37K</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AALF__gemma-2-27b-it-SimPO-37K-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AALF/gemma-2-27b-it-SimPO-37K
|
27f15219df2000a16955c9403c3f38b5f3413b3d
| 9.298079 |
gemma
| 18 | 27 | true | false | false | true | 9.997722 | 0.240653 | 24.065258 | 0.391134 | 15.307881 | 0 | 0 | 0.280201 | 4.026846 | 0.34876 | 1.595052 | 0.197141 | 10.79344 | false | false |
2024-08-13
|
2024-09-05
| 2 |
google/gemma-2-27b
|
AALF_gemma-2-27b-it-SimPO-37K-100steps_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AALF/gemma-2-27b-it-SimPO-37K-100steps" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AALF/gemma-2-27b-it-SimPO-37K-100steps</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AALF__gemma-2-27b-it-SimPO-37K-100steps-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AALF/gemma-2-27b-it-SimPO-37K-100steps
|
d5cbf18b2eb90b77f5ddbb74cfcaeedfa692c90c
| 9.894336 |
gemma
| 11 | 27 | true | false | false | true | 9.856735 | 0.256764 | 25.676427 | 0.393082 | 15.261078 | 0 | 0 | 0.288591 | 5.145414 | 0.332917 | 0.78125 | 0.212517 | 12.501847 | false | false |
2024-08-13
|
2024-09-21
| 2 |
google/gemma-2-27b
|
AELLM_gemma-2-aeria-infinity-9b_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AELLM/gemma-2-aeria-infinity-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AELLM/gemma-2-aeria-infinity-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AELLM__gemma-2-aeria-infinity-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AELLM/gemma-2-aeria-infinity-9b
|
24e1de07258925d5ddb52134b66e2eb0d698dc11
| 28.344029 | 1 | 9 | false | false | false | true | 3.003789 | 0.7594 | 75.93995 | 0.598334 | 42.090214 | 0 | 0 | 0.333893 | 11.185682 | 0.401969 | 9.046094 | 0.38622 | 31.802231 | false | false |
2024-10-09
|
2024-10-09
| 1 |
AELLM/gemma-2-aeria-infinity-9b (Merge)
|
|
AELLM_gemma-2-lyco-infinity-9b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AELLM/gemma-2-lyco-infinity-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AELLM/gemma-2-lyco-infinity-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AELLM__gemma-2-lyco-infinity-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AELLM/gemma-2-lyco-infinity-9b
|
2941a682fcbcfea3f1485c9e0691cc1d9edc742e
| 27.204937 | 0 | 10 | false | false | false | true | 2.97852 | 0.731648 | 73.164758 | 0.583953 | 39.787539 | 0 | 0 | 0.32802 | 10.402685 | 0.400635 | 8.91276 | 0.378657 | 30.961879 | false | false |
2024-10-09
|
2024-10-09
| 1 |
AELLM/gemma-2-lyco-infinity-9b (Merge)
|
|
AGI-0_Artificium-llama3.1-8B-001_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AGI-0/Artificium-llama3.1-8B-001" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AGI-0/Artificium-llama3.1-8B-001</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AGI-0__Artificium-llama3.1-8B-001-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AGI-0/Artificium-llama3.1-8B-001
|
6bf3dcca3b75a06a4e04e5f944e709cccf4673fd
| 19.063822 |
unknown
| 33 | 8 | true | false | false | true | 1.860327 | 0.524769 | 52.476872 | 0.425622 | 19.348898 | 0.110272 | 11.02719 | 0.26594 | 2.12528 | 0.379458 | 5.165625 | 0.318152 | 24.239066 | false | false |
2024-08-16
|
2024-09-08
| 0 |
AGI-0/Artificium-llama3.1-8B-001
|
AGI-0_smartllama3.1-8B-001_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AGI-0/smartllama3.1-8B-001" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AGI-0/smartllama3.1-8B-001</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AGI-0__smartllama3.1-8B-001-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AGI-0/smartllama3.1-8B-001
|
974d5ee685f1be003a1d8d08e907fe672d225035
| 20.23573 |
unknown
| 33 | 8 | true | false | false | false | 0.718834 | 0.351787 | 35.178659 | 0.467018 | 24.857737 | 0.11858 | 11.858006 | 0.306208 | 7.494407 | 0.438646 | 14.397396 | 0.348654 | 27.628177 | false | false |
2024-08-16
|
2024-11-25
| 0 |
AGI-0/smartllama3.1-8B-001
|
AI-MO_NuminaMath-7B-CoT_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AI-MO/NuminaMath-7B-CoT" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AI-MO/NuminaMath-7B-CoT</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AI-MO__NuminaMath-7B-CoT-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AI-MO/NuminaMath-7B-CoT
|
ff7e3044218efe64128bd9c21f9ec66c3de04324
| 13.097309 |
apache-2.0
| 18 | 6 | true | false | false | true | 0.745989 | 0.268854 | 26.885442 | 0.431419 | 19.152364 | 0.088369 | 8.836858 | 0.26594 | 2.12528 | 0.330344 | 0.826302 | 0.286818 | 20.757609 | false | false |
2024-07-15
|
2024-09-10
| 1 |
deepseek-ai/deepseek-math-7b-base
|
AI-MO_NuminaMath-7B-TIR_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AI-MO/NuminaMath-7B-TIR" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AI-MO/NuminaMath-7B-TIR</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AI-MO__NuminaMath-7B-TIR-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AI-MO/NuminaMath-7B-TIR
|
c6e394cc0579423c9cde6df6cc192c07dae73388
| 11.815723 |
apache-2.0
| 321 | 6 | true | false | false | false | 1.07411 | 0.275624 | 27.562423 | 0.414369 | 16.873547 | 0.018882 | 1.888218 | 0.258389 | 1.118568 | 0.350927 | 4.199219 | 0.273271 | 19.252364 | false | false |
2024-07-04
|
2024-07-11
| 1 |
deepseek-ai/deepseek-math-7b-base
|
AI-Sweden-Models_Llama-3-8B-instruct_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AI-Sweden-Models/Llama-3-8B-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AI-Sweden-Models/Llama-3-8B-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AI-Sweden-Models__Llama-3-8B-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AI-Sweden-Models/Llama-3-8B-instruct
|
4e1c955228bdb4d69c1c4560e8d5872312a8f033
| 13.777204 |
llama3
| 10 | 8 | true | false | false | true | 1.166111 | 0.240128 | 24.012841 | 0.417346 | 18.388096 | 0.004532 | 0.453172 | 0.26594 | 2.12528 | 0.477094 | 19.936719 | 0.259724 | 17.747119 | false | false |
2024-06-01
|
2024-06-27
| 2 |
meta-llama/Meta-Llama-3-8B
|
AI-Sweden-Models_gpt-sw3-40b_float16
|
float16
|
🟢 pretrained
|
🟢
|
Original
|
GPT2LMHeadModel
|
<a target="_blank" href="https://huggingface.co/AI-Sweden-Models/gpt-sw3-40b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AI-Sweden-Models/gpt-sw3-40b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AI-Sweden-Models__gpt-sw3-40b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AI-Sweden-Models/gpt-sw3-40b
|
1af27994df1287a7fac1b10d60e40ca43a22a385
| 4.734433 |
other
| 10 | 39 | true | false | false | false | 2.959819 | 0.14703 | 14.702988 | 0.326774 | 6.894934 | 0.009063 | 0.906344 | 0.234899 | 0 | 0.36324 | 2.838281 | 0.127576 | 3.064051 | false | false |
2023-02-22
|
2024-06-26
| 0 |
AI-Sweden-Models/gpt-sw3-40b
|
AbacusResearch_Jallabi-34B_float16
|
float16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/AbacusResearch/Jallabi-34B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AbacusResearch/Jallabi-34B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AbacusResearch__Jallabi-34B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AbacusResearch/Jallabi-34B
|
f65696da4ed82c9a20e94b200d9dccffa07af682
| 25.972084 |
apache-2.0
| 2 | 34 | true | false | false | false | 3.286492 | 0.35286 | 35.286041 | 0.602338 | 43.615765 | 0.039275 | 3.927492 | 0.338926 | 11.856823 | 0.482177 | 20.238802 | 0.468168 | 40.90758 | false | false |
2024-03-01
|
2024-06-27
| 0 |
AbacusResearch/Jallabi-34B
|
Alibaba-NLP_gte-Qwen2-7B-instruct_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/Alibaba-NLP/gte-Qwen2-7B-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Alibaba-NLP/gte-Qwen2-7B-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Alibaba-NLP__gte-Qwen2-7B-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Alibaba-NLP/gte-Qwen2-7B-instruct
|
e26182b2122f4435e8b3ebecbf363990f409b45b
| 13.40618 |
apache-2.0
| 226 | 7 | true | false | false | true | 2.172113 | 0.22554 | 22.554045 | 0.449514 | 21.925482 | 0.03852 | 3.851964 | 0.244966 | 0 | 0.355854 | 6.315104 | 0.332114 | 25.790485 | false | false |
2024-06-15
|
2024-08-05
| 0 |
Alibaba-NLP/gte-Qwen2-7B-instruct
|
ArliAI_ArliAI-RPMax-12B-v1.1_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/ArliAI/ArliAI-RPMax-12B-v1.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ArliAI/ArliAI-RPMax-12B-v1.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ArliAI__ArliAI-RPMax-12B-v1.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
ArliAI/ArliAI-RPMax-12B-v1.1
|
645db1cf8ad952eb57854a133e8e15303b898b04
| 20.812694 |
apache-2.0
| 41 | 12 | true | false | false | true | 1.833402 | 0.534885 | 53.488522 | 0.475182 | 24.809063 | 0.102719 | 10.271903 | 0.281879 | 4.250559 | 0.361844 | 5.563802 | 0.338431 | 26.492317 | false | false |
2024-08-31
|
2024-09-05
| 0 |
ArliAI/ArliAI-RPMax-12B-v1.1
|
ArliAI_Llama-3.1-8B-ArliAI-RPMax-v1.1_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/ArliAI__Llama-3.1-8B-ArliAI-RPMax-v1.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
|
540bd352e59c63900af91b95a932b33aaee70c76
| 23.916967 |
llama3
| 29 | 8 | true | false | false | true | 0.892745 | 0.635902 | 63.590163 | 0.501561 | 28.787014 | 0.129909 | 12.990937 | 0.283557 | 4.474273 | 0.357688 | 5.310938 | 0.355136 | 28.348478 | false | false |
2024-08-23
|
2024-09-19
| 0 |
ArliAI/Llama-3.1-8B-ArliAI-RPMax-v1.1
|
Artples_L-MChat-7b_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/Artples/L-MChat-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Artples/L-MChat-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Artples__L-MChat-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Artples/L-MChat-7b
|
e10137f5cbfc1b73068d6473e4a87241cca0b3f4
| 21.225905 |
apache-2.0
| 2 | 7 | true | false | false | true | 0.592226 | 0.529665 | 52.966462 | 0.460033 | 24.201557 | 0.09139 | 9.138973 | 0.305369 | 7.38255 | 0.402865 | 8.12474 | 0.32987 | 25.54115 | true | false |
2024-04-02
|
2024-07-07
| 1 |
Artples/L-MChat-7b (Merge)
|
Artples_L-MChat-Small_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
PhiForCausalLM
|
<a target="_blank" href="https://huggingface.co/Artples/L-MChat-Small" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Artples/L-MChat-Small</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Artples__L-MChat-Small-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Artples/L-MChat-Small
|
52484c277f6062c12dc6d6b6397ee0d0c21b0126
| 14.891449 |
mit
| 1 | 2 | true | false | false | true | 0.465511 | 0.328706 | 32.870561 | 0.482256 | 26.856516 | 0.017372 | 1.73716 | 0.267617 | 2.348993 | 0.369594 | 9.265885 | 0.246426 | 16.269577 | true | false |
2024-04-11
|
2024-07-07
| 1 |
Artples/L-MChat-Small (Merge)
|
Aryanne_SuperHeart_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Aryanne/SuperHeart" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Aryanne/SuperHeart</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Aryanne__SuperHeart-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Aryanne/SuperHeart
|
02b5050d7e600ce3db81a19638f6043c895d60cf
| 25.267673 |
llama3.1
| 1 | 8 | true | false | false | false | 0.903959 | 0.519223 | 51.922344 | 0.521538 | 31.893554 | 0.138973 | 13.897281 | 0.301174 | 6.823266 | 0.443573 | 14.713281 | 0.391207 | 32.356309 | true | false |
2024-09-23
|
2024-09-23
| 1 |
Aryanne/SuperHeart (Merge)
|
AtAndDev_Qwen2.5-1.5B-continuous-learnt_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AtAndDev/Qwen2.5-1.5B-continuous-learnt" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AtAndDev/Qwen2.5-1.5B-continuous-learnt</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AtAndDev__Qwen2.5-1.5B-continuous-learnt-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AtAndDev/Qwen2.5-1.5B-continuous-learnt
|
01c0981db9cf0f146fe050065f17343af75a8aa6
| 16.518524 | 0 | 1 | false | false | false | true | 0.673035 | 0.460521 | 46.052142 | 0.425775 | 19.537666 | 0.074773 | 7.477341 | 0.26594 | 2.12528 | 0.363646 | 3.789063 | 0.281167 | 20.129654 | false | false |
2024-10-13
| 0 |
Removed
|
||
AtAndDev_Qwen2.5-1.5B-continuous-learnt_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/AtAndDev/Qwen2.5-1.5B-continuous-learnt" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">AtAndDev/Qwen2.5-1.5B-continuous-learnt</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/AtAndDev__Qwen2.5-1.5B-continuous-learnt-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
AtAndDev/Qwen2.5-1.5B-continuous-learnt
|
01c0981db9cf0f146fe050065f17343af75a8aa6
| 16.45133 | 0 | 1 | false | false | false | true | 0.688585 | 0.451054 | 45.105431 | 0.42747 | 19.766409 | 0.085347 | 8.534743 | 0.270134 | 2.684564 | 0.362281 | 2.551823 | 0.280585 | 20.065012 | false | false |
2024-10-18
| 0 |
Removed
|
||
Aurel9_testmerge-7b_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/Aurel9/testmerge-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Aurel9/testmerge-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Aurel9__testmerge-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Aurel9/testmerge-7b
|
b5f0a72d981b5b2c6bd6294093c6956d88477a3e
| 20.994478 | 0 | 7 | false | false | false | false | 0.476464 | 0.397998 | 39.799842 | 0.518959 | 32.792793 | 0.067221 | 6.722054 | 0.300336 | 6.711409 | 0.465865 | 17.133073 | 0.305269 | 22.807698 | false | false |
2024-11-16
|
2024-11-16
| 1 |
Aurel9/testmerge-7b (Merge)
|
|
Azure99_blossom-v5-32b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/Azure99/blossom-v5-32b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Azure99/blossom-v5-32b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Azure99__blossom-v5-32b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Azure99/blossom-v5-32b
|
ccd4d86e3de01187043683dea1e28df904f7408e
| 26.352555 |
apache-2.0
| 4 | 32 | true | false | false | true | 5.688 | 0.523544 | 52.35442 | 0.595455 | 42.883056 | 0.10423 | 10.422961 | 0.311242 | 8.165548 | 0.402 | 8.35 | 0.423454 | 35.939347 | false | false |
2024-04-29
|
2024-09-21
| 0 |
Azure99/blossom-v5-32b
|
Azure99_blossom-v5-llama3-8b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Azure99/blossom-v5-llama3-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Azure99/blossom-v5-llama3-8b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Azure99__blossom-v5-llama3-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Azure99/blossom-v5-llama3-8b
|
91ea35e2e65516988021e4bb3b908e3e497e05c2
| 14.473082 |
apache-2.0
| 4 | 8 | true | false | false | true | 0.872153 | 0.434293 | 43.429323 | 0.418491 | 18.306535 | 0.043807 | 4.380665 | 0.265101 | 2.013423 | 0.367021 | 5.310938 | 0.220578 | 13.397606 | false | false |
2024-04-20
|
2024-09-21
| 0 |
Azure99/blossom-v5-llama3-8b
|
Azure99_blossom-v5.1-34b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Azure99/blossom-v5.1-34b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Azure99/blossom-v5.1-34b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Azure99__blossom-v5.1-34b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Azure99/blossom-v5.1-34b
|
2c803204f5dbf4ce37e2df98eb0205cdc53de10d
| 28.599286 |
apache-2.0
| 5 | 34 | true | false | false | true | 9.591483 | 0.569656 | 56.965629 | 0.610911 | 44.147705 | 0.1571 | 15.70997 | 0.309564 | 7.941834 | 0.392792 | 7.298958 | 0.455785 | 39.531619 | false | false |
2024-05-19
|
2024-07-27
| 0 |
Azure99/blossom-v5.1-34b
|
Azure99_blossom-v5.1-9b_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Azure99/blossom-v5.1-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Azure99/blossom-v5.1-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Azure99__blossom-v5.1-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Azure99/blossom-v5.1-9b
|
6044a3dc1e04529fe883aa513d37f266a320d793
| 24.871504 |
apache-2.0
| 2 | 8 | true | false | false | true | 2.21572 | 0.508582 | 50.858167 | 0.534329 | 34.201244 | 0.116314 | 11.63142 | 0.33557 | 11.409396 | 0.399396 | 8.024479 | 0.397939 | 33.104314 | false | false |
2024-05-15
|
2024-07-24
| 0 |
Azure99/blossom-v5.1-9b
|
BAAI_Gemma2-9B-IT-Simpo-Infinity-Preference_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Gemma2-9B-IT-Simpo-Infinity-Preference-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Gemma2-9B-IT-Simpo-Infinity-Preference
|
028a91b1a4f14d365c6db08093b03348455c7bad
| 20.984069 | 14 | 9 | false | false | false | true | 5.86346 | 0.317638 | 31.763831 | 0.597946 | 42.190844 | 0 | 0 | 0.339765 | 11.96868 | 0.396573 | 8.104948 | 0.386885 | 31.876108 | false | false |
2024-08-28
|
2024-09-05
| 2 |
google/gemma-2-9b
|
|
BAAI_Infinity-Instruct-3M-0613-Llama3-70B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0613-Llama3-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0613-Llama3-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0613-Llama3-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0613-Llama3-70B
|
9fc53668064bdda22975ca72c5a287f8241c95b3
| 34.697075 |
apache-2.0
| 5 | 70 | true | false | false | true | 10.526907 | 0.682113 | 68.211346 | 0.664161 | 51.327161 | 0.162387 | 16.238671 | 0.358221 | 14.42953 | 0.45226 | 16.532552 | 0.472989 | 41.443189 | false | false |
2024-06-27
|
2024-06-28
| 0 |
BAAI/Infinity-Instruct-3M-0613-Llama3-70B
|
BAAI_Infinity-Instruct-3M-0613-Mistral-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0613-Mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0613-Mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0613-Mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0613-Mistral-7B
|
c7a742e539ec264b9eaeefe2aed29e92e8a7ebd6
| 22.180237 |
apache-2.0
| 11 | 7 | true | false | false | true | 0.949375 | 0.531987 | 53.198735 | 0.495823 | 28.992936 | 0.074773 | 7.477341 | 0.296141 | 6.152125 | 0.435083 | 13.252083 | 0.316074 | 24.0082 | false | false |
2024-06-21
|
2024-06-27
| 0 |
BAAI/Infinity-Instruct-3M-0613-Mistral-7B
|
BAAI_Infinity-Instruct-3M-0625-Llama3-70B_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Llama3-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Llama3-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Llama3-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Llama3-70B
|
6d8ceada57e55cff3503191adc4d6379ff321fe2
| 36.142217 |
apache-2.0
| 3 | 70 | true | false | false | true | 10.430955 | 0.744212 | 74.421202 | 0.667034 | 52.028162 | 0.179003 | 17.900302 | 0.357383 | 14.317673 | 0.461656 | 18.340365 | 0.45861 | 39.845597 | false | false |
2024-07-09
|
2024-08-30
| 0 |
BAAI/Infinity-Instruct-3M-0625-Llama3-70B
|
BAAI_Infinity-Instruct-3M-0625-Llama3-8B_float16
|
float16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Llama3-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Llama3-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Llama3-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Llama3-8B
|
7be7c0ff1e35c3bb781c47222da99a1724f5f1da
| 21.60936 |
apache-2.0
| 3 | 8 | true | false | false | true | 0.858004 | 0.605027 | 60.502688 | 0.495499 | 28.988222 | 0.061178 | 6.117825 | 0.275168 | 3.355705 | 0.371208 | 5.667708 | 0.325216 | 25.02401 | false | false |
2024-07-09
|
2024-07-13
| 0 |
BAAI/Infinity-Instruct-3M-0625-Llama3-8B
|
BAAI_Infinity-Instruct-3M-0625-Mistral-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Mistral-7B
|
302e3ae0bcc50dae3fb69fc1b08b518398e8c407
| 22.843425 |
apache-2.0
| 3 | 7 | true | false | false | true | 0.785797 | 0.586742 | 58.674207 | 0.493967 | 28.823289 | 0.076284 | 7.628399 | 0.286913 | 4.9217 | 0.42724 | 12.238281 | 0.322972 | 24.774675 | false | false |
2024-07-09
|
2024-08-05
| 0 |
BAAI/Infinity-Instruct-3M-0625-Mistral-7B
|
BAAI_Infinity-Instruct-3M-0625-Qwen2-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Qwen2-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Qwen2-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Qwen2-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Qwen2-7B
|
503c24156d7682458686a7b5324f7f886e63470d
| 24.135357 |
apache-2.0
| 8 | 7 | true | false | false | true | 1.330078 | 0.555393 | 55.539302 | 0.534591 | 34.656829 | 0.068731 | 6.873112 | 0.312919 | 8.389262 | 0.38876 | 6.461719 | 0.396027 | 32.891918 | false | false |
2024-07-09
|
2024-08-05
| 0 |
BAAI/Infinity-Instruct-3M-0625-Qwen2-7B
|
BAAI_Infinity-Instruct-3M-0625-Yi-1.5-9B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-3M-0625-Yi-1.5-9B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B
|
a42c86c61b98ca4fdf238d688fe6ea11cf414d29
| 27.943551 |
apache-2.0
| 3 | 8 | true | false | false | true | 1.116801 | 0.518598 | 51.859843 | 0.550912 | 35.378707 | 0.151813 | 15.181269 | 0.354027 | 13.870246 | 0.457531 | 16.72474 | 0.411818 | 34.646498 | false | false |
2024-07-09
|
2024-08-05
| 0 |
BAAI/Infinity-Instruct-3M-0625-Yi-1.5-9B
|
BAAI_Infinity-Instruct-7M-0729-Llama3_1-8B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-0729-Llama3_1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-0729-Llama3_1-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-0729-Llama3_1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-0729-Llama3_1-8B
|
0aca33fd7500a781d041e8bf7e5e3789b03f54f4
| 23.094956 |
llama3.1
| 8 | 8 | true | false | false | true | 0.866805 | 0.613195 | 61.319521 | 0.507734 | 30.888805 | 0.106495 | 10.649547 | 0.292785 | 5.704698 | 0.357844 | 5.297135 | 0.32239 | 24.710033 | false | false |
2024-08-02
|
2024-08-05
| 0 |
BAAI/Infinity-Instruct-7M-0729-Llama3_1-8B
|
BAAI_Infinity-Instruct-7M-0729-mistral-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-0729-mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-0729-mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-0729-mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-0729-mistral-7B
|
36651591cb13346ecbde23832013e024029700fa
| 22.914335 |
apache-2.0
| 3 | 7 | true | false | false | true | 0.799261 | 0.616193 | 61.619281 | 0.496381 | 28.697915 | 0.064955 | 6.495468 | 0.290268 | 5.369128 | 0.406188 | 10.040104 | 0.327377 | 25.264111 | false | false |
2024-07-25
|
2024-08-05
| 0 |
BAAI/Infinity-Instruct-7M-0729-mistral-7B
|
BAAI_Infinity-Instruct-7M-Gen-Llama3_1-70B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-Gen-Llama3_1-70B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B
|
1ef63c4993a8c723c9695c827295c17080a64435
| 37.10681 |
llama3.1
| 17 | 70 | true | false | false | true | 11.069121 | 0.733546 | 73.354588 | 0.66952 | 52.498947 | 0.229607 | 22.960725 | 0.375839 | 16.778523 | 0.453906 | 16.971615 | 0.460688 | 40.076463 | false | false |
2024-07-25
|
2024-09-26
| 0 |
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-70B
|
BAAI_Infinity-Instruct-7M-Gen-Llama3_1-8B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-Gen-Llama3_1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-Gen-Llama3_1-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-Gen-Llama3_1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-8B
|
56f9c2845ae024eb8b1dd9ea0d8891cbaf33c596
| 23.094956 |
llama3.1
| 8 | 8 | true | false | false | true | 0.91714 | 0.613195 | 61.319521 | 0.507734 | 30.888805 | 0.106495 | 10.649547 | 0.292785 | 5.704698 | 0.357844 | 5.297135 | 0.32239 | 24.710033 | false | false |
2024-08-02
|
2024-08-29
| 0 |
BAAI/Infinity-Instruct-7M-Gen-Llama3_1-8B
|
BAAI_Infinity-Instruct-7M-Gen-mistral-7B_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/Infinity-Instruct-7M-Gen-mistral-7B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/Infinity-Instruct-7M-Gen-mistral-7B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__Infinity-Instruct-7M-Gen-mistral-7B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/Infinity-Instruct-7M-Gen-mistral-7B
|
82c83d670a8954f4250547b53a057dea1fbd460d
| 22.888939 |
apache-2.0
| 3 | 7 | true | false | false | true | 0.824635 | 0.614669 | 61.466908 | 0.496381 | 28.697915 | 0.064955 | 6.495468 | 0.290268 | 5.369128 | 0.406188 | 10.040104 | 0.327377 | 25.264111 | false | false |
2024-07-25
|
2024-08-29
| 0 |
BAAI/Infinity-Instruct-7M-Gen-mistral-7B
|
BAAI_OPI-Llama-3.1-8B-Instruct_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BAAI/OPI-Llama-3.1-8B-Instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BAAI/OPI-Llama-3.1-8B-Instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BAAI__OPI-Llama-3.1-8B-Instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BAAI/OPI-Llama-3.1-8B-Instruct
|
48504799d009b4e1b29e6d2948a7cde68acdc3b0
| 8.305018 |
llama3.1
| 1 | 8 | true | false | false | true | 0.671657 | 0.207455 | 20.745511 | 0.355122 | 9.768712 | 0 | 0 | 0.274329 | 3.243848 | 0.323302 | 3.579427 | 0.212434 | 12.492612 | false | false |
2024-09-06
|
2024-09-21
| 2 |
meta-llama/Meta-Llama-3.1-8B
|
BEE-spoke-data_Meta-Llama-3-8Bee_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/Meta-Llama-3-8Bee" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/Meta-Llama-3-8Bee</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__Meta-Llama-3-8Bee-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/Meta-Llama-3-8Bee
|
8143e34e77a49a30ec2617c5c9cc22cb3cda2287
| 14.544519 |
llama3
| 0 | 8 | true | false | false | false | 0.83038 | 0.195066 | 19.506576 | 0.462636 | 24.199033 | 0.041541 | 4.154079 | 0.313758 | 8.501119 | 0.365406 | 6.242448 | 0.321975 | 24.663859 | false | false |
2024-04-28
|
2024-07-04
| 1 |
meta-llama/Meta-Llama-3-8B
|
BEE-spoke-data_smol_llama-101M-GQA_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/smol_llama-101M-GQA" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/smol_llama-101M-GQA</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__smol_llama-101M-GQA-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/smol_llama-101M-GQA
|
bb26643db413bada7e0c3c50752bf9da82403dba
| 3.918895 |
apache-2.0
| 28 | 0 | true | false | false | false | 0.119606 | 0.138437 | 13.843712 | 0.301756 | 3.198004 | 0 | 0 | 0.25755 | 1.006711 | 0.371271 | 4.275521 | 0.110705 | 1.189421 | false | false |
2023-10-26
|
2024-07-06
| 0 |
BEE-spoke-data/smol_llama-101M-GQA
|
BEE-spoke-data_smol_llama-220M-GQA_bfloat16
|
bfloat16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/smol_llama-220M-GQA" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/smol_llama-220M-GQA</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__smol_llama-220M-GQA-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/smol_llama-220M-GQA
|
8845b1d3c0bc73522ef2700aab467183cbdca9f7
| 6.401567 |
apache-2.0
| 12 | 0 | true | false | false | false | 0.163613 | 0.238605 | 23.860468 | 0.303167 | 3.037843 | 0 | 0 | 0.255872 | 0.782998 | 0.405875 | 9.067708 | 0.114943 | 1.660387 | false | false |
2023-12-22
|
2024-06-26
| 0 |
BEE-spoke-data/smol_llama-220M-GQA
|
BEE-spoke-data_smol_llama-220M-GQA-fineweb_edu_bfloat16
|
bfloat16
|
🟩 continuously pretrained
|
🟩
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__smol_llama-220M-GQA-fineweb_edu-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/smol_llama-220M-GQA-fineweb_edu
|
dec16b41d5e94070dbc1f8449a554373fd4cc1d1
| 6.516558 |
apache-2.0
| 1 | 0 | true | false | false | false | 0.161876 | 0.198812 | 19.881248 | 0.292905 | 2.314902 | 0 | 0 | 0.259228 | 1.230425 | 0.43676 | 14.261719 | 0.112699 | 1.411052 | false | false |
2024-06-08
|
2024-06-26
| 1 |
BEE-spoke-data/smol_llama-220M-GQA
|
BEE-spoke-data_smol_llama-220M-openhermes_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/smol_llama-220M-openhermes" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/smol_llama-220M-openhermes</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__smol_llama-220M-openhermes-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/smol_llama-220M-openhermes
|
fb4bcd4b7eee363baacb4176a26cea2aaeb173f4
| 4.761772 |
apache-2.0
| 5 | 0 | true | false | false | false | 0.154426 | 0.155523 | 15.55229 | 0.302752 | 3.107692 | 0 | 0 | 0.267617 | 2.348993 | 0.384729 | 6.224479 | 0.112035 | 1.337175 | false | false |
2023-12-30
|
2024-09-21
| 1 |
BEE-spoke-data/smol_llama-220M-GQA
|
BEE-spoke-data_tFINE-900m-e16-d32-flan_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
T5ForConditionalGeneration
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-flan" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/tFINE-900m-e16-d32-flan</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__tFINE-900m-e16-d32-flan-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/tFINE-900m-e16-d32-flan
|
d9ffec9798402d13d8f2c56ec3de3ad092445297
| 4.433887 |
apache-2.0
| 0 | 0 | true | false | false | false | 2.456006 | 0.150577 | 15.057714 | 0.302804 | 4.411894 | 0 | 0 | 0.233221 | 0 | 0.372417 | 3.71875 | 0.130735 | 3.414967 | false | false |
2024-09-06
|
2024-09-13
| 1 |
pszemraj/tFINE-900m-e16-d32-1024ctx
|
BEE-spoke-data_tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
T5ForConditionalGeneration
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/tFINE-900m-e16-d32-flan-infinity-instruct-7m-T2T_en-1024
|
b1e2f12f5224be9f7da0cb5ff30e1bbb3f10f6ca
| 5.823653 |
apache-2.0
| 0 | 0 | true | false | false | false | 2.600608 | 0.132067 | 13.206736 | 0.313779 | 4.737018 | 0 | 0 | 0.254195 | 0.559284 | 0.439271 | 13.808854 | 0.12367 | 2.630024 | false | false |
2024-09-10
|
2024-09-14
| 2 |
pszemraj/tFINE-900m-e16-d32-1024ctx
|
BEE-spoke-data_tFINE-900m-e16-d32-instruct_2e_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
T5ForConditionalGeneration
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/tFINE-900m-e16-d32-instruct_2e" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/tFINE-900m-e16-d32-instruct_2e</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__tFINE-900m-e16-d32-instruct_2e-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/tFINE-900m-e16-d32-instruct_2e
|
4c626138c9f4e0c3eafe74b2755eb89334c7ca59
| 5.681552 |
apache-2.0
| 0 | 0 | true | false | false | false | 2.516619 | 0.140286 | 14.028555 | 0.313457 | 5.01307 | 0 | 0 | 0.259228 | 1.230425 | 0.420698 | 11.18724 | 0.12367 | 2.630024 | false | false |
2024-09-17
|
2024-09-22
| 3 |
pszemraj/tFINE-900m-e16-d32-1024ctx
|
BEE-spoke-data_tFINE-900m-instruct-orpo_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
T5ForConditionalGeneration
|
<a target="_blank" href="https://huggingface.co/BEE-spoke-data/tFINE-900m-instruct-orpo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BEE-spoke-data/tFINE-900m-instruct-orpo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BEE-spoke-data__tFINE-900m-instruct-orpo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BEE-spoke-data/tFINE-900m-instruct-orpo
|
e0a21c79bac74442252d36e2c01403afa3f0971b
| 3.431957 |
apache-2.0
| 0 | 0 | true | false | false | true | 2.574962 | 0.132992 | 13.299157 | 0.302209 | 3.267301 | 0 | 0 | 0.259228 | 1.230425 | 0.340854 | 1.106771 | 0.115193 | 1.688091 | false | false |
2024-09-22
|
2024-09-23
| 0 |
BEE-spoke-data/tFINE-900m-instruct-orpo
|
BSC-LT_salamandra-7b_float16
|
float16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BSC-LT/salamandra-7b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BSC-LT/salamandra-7b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BSC-LT__salamandra-7b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BSC-LT/salamandra-7b
|
bf30739316ceac4b624583a27ec96dfc401179e8
| 5.641971 |
apache-2.0
| 13 | 7 | true | false | false | false | 0.189289 | 0.136738 | 13.67383 | 0.351661 | 10.157422 | 0 | 0 | 0.270134 | 2.684564 | 0.350094 | 1.861719 | 0.149269 | 5.474291 | false | false |
2024-09-30
|
2024-11-22
| 0 |
BSC-LT/salamandra-7b
|
BSC-LT_salamandra-7b-instruct_float16
|
float16
|
🟢 pretrained
|
🟢
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BSC-LT/salamandra-7b-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BSC-LT/salamandra-7b-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BSC-LT__salamandra-7b-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BSC-LT/salamandra-7b-instruct
|
77ddccbc7d9f9ffd55a8535365e8eebc493ccb8e
| 10.080539 |
apache-2.0
| 32 | 7 | true | false | false | true | 1.147504 | 0.245074 | 24.507418 | 0.385132 | 14.688129 | 0.002266 | 0.226586 | 0.264262 | 1.901566 | 0.413437 | 10.213021 | 0.180519 | 8.946513 | false | false |
2024-09-30
|
2024-11-22
| 1 |
BSC-LT/salamandra-7b-instruct (Merge)
|
Ba2han_Llama-Phi-3_DoRA_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/Ba2han/Llama-Phi-3_DoRA" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Ba2han/Llama-Phi-3_DoRA</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Ba2han__Llama-Phi-3_DoRA-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Ba2han/Llama-Phi-3_DoRA
|
36f99064a7be8ba475c2ee5c5424e95c263ccb87
| 25.318838 |
mit
| 6 | 3 | true | false | false | true | 0.533136 | 0.513053 | 51.305314 | 0.551456 | 37.249164 | 0.112538 | 11.253776 | 0.326342 | 10.178971 | 0.406927 | 9.532552 | 0.391539 | 32.393248 | false | false |
2024-05-15
|
2024-06-26
| 0 |
Ba2han/Llama-Phi-3_DoRA
|
BenevolenceMessiah_Qwen2.5-72B-2x-Instruct-TIES-v1.0_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
Qwen2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BenevolenceMessiah__Qwen2.5-72B-2x-Instruct-TIES-v1.0-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0
|
459891ec78c9bbed2836a8bba706e1707db10231
| 34.185749 | 0 | 72 | false | false | false | true | 17.350892 | 0.54735 | 54.734992 | 0.727311 | 61.911495 | 0.093656 | 9.365559 | 0.36745 | 15.659955 | 0.420667 | 12.016667 | 0.562832 | 51.425827 | false | false |
2024-11-11
|
2024-11-24
| 1 |
BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0 (Merge)
|
|
BenevolenceMessiah_Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
MixtralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BenevolenceMessiah__Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0
|
d90f6e36584dc9b367461701e83c833bdeb736f2
| 15.096268 |
apache-2.0
| 0 | 28 | true | true | false | false | 3.334797 | 0.301153 | 30.115316 | 0.490867 | 26.877991 | 0.043051 | 4.305136 | 0.262584 | 1.677852 | 0.407979 | 8.930729 | 0.268035 | 18.670582 | true | false |
2024-09-21
|
2024-09-22
| 1 |
BenevolenceMessiah/Yi-Coder-9B-Chat-Instruct-TIES-MoE-v1.0 (Merge)
|
BlackBeenie_Bloslain-8B-v0.2_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Bloslain-8B-v0.2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Bloslain-8B-v0.2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Bloslain-8B-v0.2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Bloslain-8B-v0.2
|
ebcb7f9f30bc172523a827d1ddefeb52b1aba494
| 23.803914 | 1 | 8 | false | false | false | false | 0.691763 | 0.502337 | 50.233713 | 0.511088 | 30.662902 | 0.145015 | 14.501511 | 0.306208 | 7.494407 | 0.407573 | 10.446615 | 0.365359 | 29.484338 | false | false |
2024-11-19
|
2024-11-19
| 1 |
BlackBeenie/Bloslain-8B-v0.2 (Merge)
|
|
BlackBeenie_Llama-3.1-8B-pythonic-passthrough-merge_float16
|
float16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Llama-3.1-8B-pythonic-passthrough-merge" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Llama-3.1-8B-pythonic-passthrough-merge</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Llama-3.1-8B-pythonic-passthrough-merge-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Llama-3.1-8B-pythonic-passthrough-merge
|
3ec46616f5b34821b3b928938931295f92e49213
| 7.311462 | 0 | 20 | false | false | false | false | 3.58329 | 0.231586 | 23.158553 | 0.345385 | 9.359905 | 0.006042 | 0.60423 | 0.268456 | 2.46085 | 0.377812 | 4.593229 | 0.133228 | 3.692007 | false | false |
2024-11-06
|
2024-11-06
| 1 |
BlackBeenie/Llama-3.1-8B-pythonic-passthrough-merge (Merge)
|
|
BlackBeenie_Neos-Gemma-2-9b_float16
|
float16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Gemma2ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Neos-Gemma-2-9b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Neos-Gemma-2-9b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Neos-Gemma-2-9b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Neos-Gemma-2-9b
|
56dbbb4f972be887e5b57311a8a32e148e98d154
| 25.211313 |
apache-2.0
| 1 | 9 | true | false | false | true | 2.679092 | 0.587567 | 58.756655 | 0.550298 | 35.638851 | 0.082326 | 8.232628 | 0.322987 | 9.731544 | 0.36175 | 5.785417 | 0.398105 | 33.122784 | false | false |
2024-11-11
|
2024-11-11
| 1 |
BlackBeenie/Neos-Gemma-2-9b (Merge)
|
BlackBeenie_Neos-Llama-3.1-8B_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Neos-Llama-3.1-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Neos-Llama-3.1-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Neos-Llama-3.1-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Neos-Llama-3.1-8B
|
9b48520ec1a777be0f1fd88f95454d85ac568407
| 19.461825 |
apache-2.0
| 1 | 8 | true | false | false | true | 0.793867 | 0.494394 | 49.439376 | 0.4425 | 21.080123 | 0.129154 | 12.915408 | 0.268456 | 2.46085 | 0.37499 | 5.740365 | 0.326213 | 25.134826 | false | false |
2024-11-12
|
2024-11-12
| 1 |
BlackBeenie/Neos-Llama-3.1-8B (Merge)
|
BlackBeenie_Neos-Llama-3.1-base_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Neos-Llama-3.1-base" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Neos-Llama-3.1-base</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Neos-Llama-3.1-base-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Neos-Llama-3.1-base
|
d4af4d73ba5fea0275fd1e3ba5102a79ac8009db
| 3.968795 | 0 | 4 | false | false | false | true | 1.409285 | 0.175082 | 17.508212 | 0.293034 | 2.221447 | 0 | 0 | 0.237416 | 0 | 0.349906 | 2.838281 | 0.111203 | 1.244829 | false | false |
2024-11-11
|
2024-11-11
| 0 |
BlackBeenie/Neos-Llama-3.1-base
|
|
BlackBeenie_Neos-Phi-3-14B-v0.1_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
Phi3ForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/Neos-Phi-3-14B-v0.1" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/Neos-Phi-3-14B-v0.1</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__Neos-Phi-3-14B-v0.1-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/Neos-Phi-3-14B-v0.1
|
0afb7cc74a94f11f2695dc92788cdc6e28325f9c
| 26.843485 |
apache-2.0
| 0 | 13 | true | false | false | true | 0.909626 | 0.402245 | 40.224493 | 0.621193 | 46.631387 | 0.166918 | 16.691843 | 0.305369 | 7.38255 | 0.412542 | 10.534375 | 0.456366 | 39.596262 | false | false |
2024-11-27
|
2024-11-27
| 1 |
BlackBeenie/Neos-Phi-3-14B-v0.1 (Merge)
|
BlackBeenie_llama-3-luminous-merged_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/llama-3-luminous-merged" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/llama-3-luminous-merged</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__llama-3-luminous-merged-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/llama-3-luminous-merged
|
64288dd8e3305f2dc11d84fe0c653f351b2e8a9d
| 21.480108 | 0 | 8 | false | false | false | false | 0.763854 | 0.432345 | 43.234507 | 0.515392 | 30.643687 | 0.07855 | 7.854985 | 0.292785 | 5.704698 | 0.414896 | 10.628646 | 0.377327 | 30.814125 | false | false |
2024-09-15
|
2024-10-11
| 1 |
BlackBeenie/llama-3-luminous-merged (Merge)
|
|
BlackBeenie_llama-3.1-8B-Galore-openassistant-guanaco_bfloat16
|
bfloat16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BlackBeenie__llama-3.1-8B-Galore-openassistant-guanaco-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco
|
828fa03c10e9085700b7abbe26f95067fab010fd
| 18.072101 | 1 | 8 | false | false | false | false | 0.85682 | 0.263484 | 26.348422 | 0.521337 | 31.444705 | 0.048338 | 4.833837 | 0.300336 | 6.711409 | 0.440625 | 14.578125 | 0.320645 | 24.516105 | false | false |
2024-10-16
|
2024-10-19
| 0 |
BlackBeenie/llama-3.1-8B-Galore-openassistant-guanaco
|
|
BoltMonkey_DreadMix_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BoltMonkey/DreadMix" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BoltMonkey/DreadMix</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BoltMonkey__DreadMix-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BoltMonkey/DreadMix
|
ab5dbaaff606538db73b6fd89aa169760104a566
| 28.661027 | 0 | 8 | false | false | false | true | 1.614205 | 0.709491 | 70.949082 | 0.54351 | 34.845015 | 0.149547 | 14.954683 | 0.299497 | 6.599553 | 0.421219 | 13.61901 | 0.378989 | 30.998818 | false | false |
2024-10-12
|
2024-10-13
| 1 |
BoltMonkey/DreadMix (Merge)
|
|
BoltMonkey_NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BoltMonkey__NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated
|
969e4c9b41e733a367f5ea18ed50a6171b5e2357
| 27.726282 |
llama3.1
| 2 | 8 | true | false | false | true | 1.640513 | 0.799891 | 79.989096 | 0.515199 | 30.7599 | 0.116314 | 11.63142 | 0.28104 | 4.138702 | 0.401875 | 9.467708 | 0.373338 | 30.370863 | true | false |
2024-10-01
|
2024-10-10
| 1 |
BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated (Merge)
|
BoltMonkey_NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BoltMonkey__NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated
|
969e4c9b41e733a367f5ea18ed50a6171b5e2357
| 21.345511 |
llama3.1
| 2 | 8 | true | false | false | false | 0.774319 | 0.459023 | 45.902317 | 0.518544 | 30.793785 | 0.093656 | 9.365559 | 0.274329 | 3.243848 | 0.40826 | 9.532552 | 0.363115 | 29.235003 | true | false |
2024-10-01
|
2024-10-01
| 1 |
BoltMonkey/NeuralDaredevil-SuperNova-Lite-7B-DARETIES-abliterated (Merge)
|
BoltMonkey_SuperNeuralDreadDevil-8b_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BoltMonkey/SuperNeuralDreadDevil-8b" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BoltMonkey/SuperNeuralDreadDevil-8b</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BoltMonkey__SuperNeuralDreadDevil-8b-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BoltMonkey/SuperNeuralDreadDevil-8b
|
804d5864127e603abec179a159b43f446246fafc
| 21.847726 | 1 | 8 | false | false | false | true | 2.405331 | 0.485801 | 48.580101 | 0.515108 | 30.606714 | 0.090634 | 9.063444 | 0.285235 | 4.697987 | 0.415948 | 10.426823 | 0.349402 | 27.711288 | false | false |
2024-10-13
|
2024-10-13
| 1 |
BoltMonkey/SuperNeuralDreadDevil-8b (Merge)
|
|
BrainWave-ML_llama3.2-3B-maths-orpo_float16
|
float16
|
🔶 fine-tuned on domain-specific datasets
|
🔶
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/BrainWave-ML/llama3.2-3B-maths-orpo" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BrainWave-ML/llama3.2-3B-maths-orpo</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BrainWave-ML__llama3.2-3B-maths-orpo-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BrainWave-ML/llama3.2-3B-maths-orpo
|
d149d83d8e8f3883421d800848fec85766181923
| 5.076083 |
apache-2.0
| 2 | 3 | true | false | false | false | 0.707219 | 0.204907 | 20.490742 | 0.291178 | 2.347041 | 0 | 0 | 0.259228 | 1.230425 | 0.357531 | 4.52474 | 0.116772 | 1.863549 | false | false |
2024-10-24
|
2024-10-24
| 2 |
meta-llama/Llama-3.2-3B-Instruct
|
BramVanroy_GEITje-7B-ultra_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
MistralForCausalLM
|
<a target="_blank" href="https://huggingface.co/BramVanroy/GEITje-7B-ultra" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BramVanroy/GEITje-7B-ultra</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BramVanroy__GEITje-7B-ultra-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BramVanroy/GEITje-7B-ultra
|
d4552cdc6f015754646464d8411aa4f6bcdba8e8
| 10.909606 |
cc-by-nc-4.0
| 37 | 7 | true | false | false | true | 0.619523 | 0.372344 | 37.234427 | 0.377616 | 12.879913 | 0.009063 | 0.906344 | 0.262584 | 1.677852 | 0.328979 | 1.522396 | 0.20113 | 11.236702 | false | false |
2024-01-27
|
2024-10-28
| 3 |
mistralai/Mistral-7B-v0.1
|
BramVanroy_fietje-2_bfloat16
|
bfloat16
|
🟩 continuously pretrained
|
🟩
|
Original
|
PhiForCausalLM
|
<a target="_blank" href="https://huggingface.co/BramVanroy/fietje-2" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BramVanroy/fietje-2</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BramVanroy__fietje-2-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BramVanroy/fietje-2
|
3abe75d01094b713368e3d911ffb78a2d66ead22
| 9.027007 |
mit
| 6 | 2 | true | false | false | false | 0.312539 | 0.209803 | 20.980332 | 0.403567 | 15.603676 | 0.009063 | 0.906344 | 0.254195 | 0.559284 | 0.369563 | 5.161979 | 0.198554 | 10.950428 | false | false |
2024-04-09
|
2024-10-28
| 1 |
microsoft/phi-2
|
BramVanroy_fietje-2-chat_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
PhiForCausalLM
|
<a target="_blank" href="https://huggingface.co/BramVanroy/fietje-2-chat" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BramVanroy/fietje-2-chat</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BramVanroy__fietje-2-chat-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BramVanroy/fietje-2-chat
|
364e785d90438b787b94e33741a930c9932353c0
| 10.388869 |
mit
| 1 | 2 | true | false | false | true | 0.399033 | 0.291736 | 29.173593 | 0.414975 | 17.718966 | 0.005287 | 0.528701 | 0.239933 | 0 | 0.35276 | 3.195052 | 0.205452 | 11.716903 | false | false |
2024-04-29
|
2024-10-28
| 3 |
microsoft/phi-2
|
BramVanroy_fietje-2-instruct_bfloat16
|
bfloat16
|
💬 chat models (RLHF, DPO, IFT, ...)
|
💬
|
Original
|
PhiForCausalLM
|
<a target="_blank" href="https://huggingface.co/BramVanroy/fietje-2-instruct" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">BramVanroy/fietje-2-instruct</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/BramVanroy__fietje-2-instruct-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
BramVanroy/fietje-2-instruct
|
b7b44797cd52eda1182667217e8371dbdfee4976
| 10.196192 |
mit
| 2 | 2 | true | false | false | true | 0.324395 | 0.278996 | 27.89964 | 0.413607 | 17.57248 | 0.005287 | 0.528701 | 0.233221 | 0 | 0.336917 | 2.914583 | 0.210356 | 12.261746 | false | false |
2024-04-27
|
2024-10-28
| 2 |
microsoft/phi-2
|
Casual-Autopsy_L3-Umbral-Mind-RP-v2.0-8B_bfloat16
|
bfloat16
|
🤝 base merges and moerges
|
🤝
|
Original
|
LlamaForCausalLM
|
<a target="_blank" href="https://huggingface.co/Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B</a> <a target="_blank" href="https://huggingface.co/datasets/open-llm-leaderboard/Casual-Autopsy__L3-Umbral-Mind-RP-v2.0-8B-details" style="color: var(--link-text-color); text-decoration: underline;text-decoration-style: dotted;">📑</a>
|
Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B
|
b46c066ea8387264858dc3461f382e7b42fd9c48
| 25.911927 |
llama3
| 12 | 8 | true | false | false | true | 0.988385 | 0.712263 | 71.226346 | 0.526241 | 32.486278 | 0.110272 | 11.02719 | 0.286913 | 4.9217 | 0.368667 | 5.55 | 0.37234 | 30.260047 | true | false |
2024-06-26
|
2024-07-02
| 1 |
Casual-Autopsy/L3-Umbral-Mind-RP-v2.0-8B (Merge)
|
End of preview. Expand
in Data Studio
README.md exists but content is empty.
- Downloads last month
- 8