amphora commited on
Commit
845c7da
·
verified ·
1 Parent(s): 2b10415

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -8
README.md CHANGED
@@ -5,15 +5,15 @@ tags: []
5
 
6
  # KO-REAson
7
 
8
- **KO-REAson** is a series of Korean-centric reasoning language models developed in collaboration with OneLineAI, KISTI, HAE-RAE and ORACLE.
9
 
10
  We use the **Language-Mixed Chain-of-Thought (CoT)** approach, which allows the model to alternate between English and Korean during the “Think” stage of reasoning, preserving key Korean terms while leveraging English for logical scaffolding.
11
 
12
- Top-performing models of our series [KO-REAson-AX3_1-7B-0831](https://huggingface.co/KoReason/KO-REASon-AX3_1-7B-0831) and [KO-REAson-7B-Q2_5-0831](https://huggingface.co/KoReason/KO-REASon-7B-Q2_5-0831) show performance comparable to models trained on closed-source datasets such as Exaone-Deep-7.8B.
13
 
14
 
15
  <p align="left">
16
- <img src="https://cdn-uploads.huggingface.co/production/uploads/60d3e619b8448e1785bbda2a/CaHtUra_lWZmg04d-QNtm.png"
17
  alt="Model Comparison" width="750"/>
18
  <br>
19
  <em style="display:inline-block; max-width:750px; text-align:cener; white-space:normal; word-wrap:break-word; line-height:1.5;">
@@ -29,9 +29,9 @@ The **KO-REAson-0831** family comes in six variants based on the base model used
29
  | Model (link) | Base | Notes |
30
  | -------------------------------------------------------------------------------------------- | -------------------- | --------------------------- |
31
  | [KO-REAson-L3_1-8B-0831](https://huggingface.co/KoReason/KO-REASon-L3_1-8B-0831) | [Llama-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) | `L3_1` → Llama-3.1-8B |
32
- | [KO-REAson-KL3_1-8B-0831](https://huggingface.co/KoReason/KO-REASon-KL3_1-8B-0831) | [Koni-Llama-3.1-8B](https://huggingface.co/KISTI-KONI/KONI-Llama3.1-8B-Instruct-20241024) | `KL3_1` → Koni-Llama-3.1-8B |
33
  | [KO-REAson-G3-4B-0831](https://huggingface.co/KoReason/KO-REASon-G3-4B-0831) | [Gemma-3 4B](https://huggingface.co/google/gemma-3-4b-it) | `G3` → Gemma-3-4B |
34
- | [KO-REAson-AX3_1-7B-0831](https://huggingface.co/KoReason/KO-REASon-AX3_1-7B-0831) | [A.X.-3.1-Light (≈7B)](https://huggingface.co/skt/A.X-3.1-Light) | `AX3_1` → A.X.-3.1-Light |
35
  | [KO-REAson-K2505_8B-0831](https://huggingface.co/KoReason/KO-REASon-K2505_8B-0831) | [Kanana-2505 (8B)](https://huggingface.co/kakaocorp/kanana-1.5-8b-instruct-2505) | `K2505` → Kanana-2505 |
36
  | [KO-REAson-7B-Q2_5-0831](https://huggingface.co/KoReason/KO-REASon-7B-Q2_5-0831) | [Qwen-2.5 (7B)](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) | `Q2_5` → Qwen-2.5 |
37
 
@@ -70,7 +70,7 @@ This separation is designed to prevent inadvertent overfitting to the benchmarks
70
  </thead>
71
  <tbody>
72
  <tr>
73
- <th>KO-REASon-AX3_1-7B-0831(Ours)</th>
74
  <td>260k</td>
75
  <td>SFT</td>
76
  <td><b>44.6</b></td>
@@ -85,6 +85,14 @@ This separation is designed to prevent inadvertent overfitting to the benchmarks
85
  <td>38.75</td>
86
  <td><u>49.95</u></td>
87
  </tr>
 
 
 
 
 
 
 
 
88
  <tr>
89
  <td colspan="6" style="text-align:center; font-weight:bold;">Open Recipe (En)</td>
90
  </tr>
@@ -162,7 +170,7 @@ This separation is designed to prevent inadvertent overfitting to the benchmarks
162
  <th>AIME 2024</th>
163
  <th>AIME 2025</th>
164
  <th>GPQA</th>
165
- <th>CLiCK</th>
166
  <th>KoBALT-700</th>
167
  </tr>
168
  </thead>
@@ -200,7 +208,11 @@ This separation is designed to prevent inadvertent overfitting to the benchmarks
200
  <td>7.61</td><td>28.46</td><td>19.31</td><td>51.61</td><td>46.67</td><td>33.33</td><td>28.28</td><td>32.48</td><td>4.71</td><td>30.61</td><td>27.31</td>
201
  </tr>
202
  <tr>
203
- <td><b>KO-REASon-AX3_1-7B-0831</b></td>
 
 
 
 
204
  <td>7.26</td><td>45.57</td><td>38.13</td><td>52.80</td><td>53.33</td><td>33.33</td><td>36.87</td><td><b>62.86</b></td><td>23.43</td><td><u>43.29</u></td><td><u>44.56</u></td>
205
  </tr>
206
  <tr>
 
5
 
6
  # KO-REAson
7
 
8
+ **KO-REAson** is a series of Korean-centric reasoning language models developed in collaboration with [OneLineAI](https://onelineai.com/), [KISTI-KONI](https://huggingface.co/KISTI-KONI), [HAE-RAE](https://huggingface.co/HAERAE-HUB) and ORACLE.
9
 
10
  We use the **Language-Mixed Chain-of-Thought (CoT)** approach, which allows the model to alternate between English and Korean during the “Think” stage of reasoning, preserving key Korean terms while leveraging English for logical scaffolding.
11
 
12
+ Top-performing models of our series [KO-REAson-AX3_1-7B-0831 (KONI-7B-R-20250831)](https://huggingface.co/KISTI-KONI/KONI-7B-R-20250831) and [KO-REAson-7B-Q2_5-0831](https://huggingface.co/KoReason/KO-REASon-7B-Q2_5-0831) show performance comparable to models trained on closed-source datasets such as Exaone-Deep-7.8B.
13
 
14
 
15
  <p align="left">
16
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/60d3e619b8448e1785bbda2a/uqrKdxbQEqAFknYBmuH7Y.png"
17
  alt="Model Comparison" width="750"/>
18
  <br>
19
  <em style="display:inline-block; max-width:750px; text-align:cener; white-space:normal; word-wrap:break-word; line-height:1.5;">
 
29
  | Model (link) | Base | Notes |
30
  | -------------------------------------------------------------------------------------------- | -------------------- | --------------------------- |
31
  | [KO-REAson-L3_1-8B-0831](https://huggingface.co/KoReason/KO-REASon-L3_1-8B-0831) | [Llama-3.1-8B](https://huggingface.co/meta-llama/Llama-3.1-8B-Instruct) | `L3_1` → Llama-3.1-8B |
32
+ | [KO-REAson-KL3_1-8B-0831](https://huggingface.co/KOREAson/KO-REAson-KL3_1-8B-0831) | [Koni-Llama-3.1-8B](https://huggingface.co/KISTI-KONI/KONI-Llama3.1-8B-Instruct-20241024) | `KL3_1` → Koni-Llama-3.1-8B; also called [KONI-Llama3.1-8B-R-20250831](https://huggingface.co/KISTI-KONI/KONI-Llama3.1-8B-R-20250831) |
33
  | [KO-REAson-G3-4B-0831](https://huggingface.co/KoReason/KO-REASon-G3-4B-0831) | [Gemma-3 4B](https://huggingface.co/google/gemma-3-4b-it) | `G3` → Gemma-3-4B |
34
+ | [KO-REAson-AX3_1-7B-0831](https://huggingface.co/KOREAson/KO-REAson-7B-AX3_1-0831) | [A.X.-3.1-Light (≈7B)](https://huggingface.co/skt/A.X-3.1-Light) | `AX3_1` → A.X.-3.1-Light; also called [KONI-7B-R-20250831](https://huggingface.co/KISTI-KONI/KONI-7B-R-20250831) |
35
  | [KO-REAson-K2505_8B-0831](https://huggingface.co/KoReason/KO-REASon-K2505_8B-0831) | [Kanana-2505 (8B)](https://huggingface.co/kakaocorp/kanana-1.5-8b-instruct-2505) | `K2505` → Kanana-2505 |
36
  | [KO-REAson-7B-Q2_5-0831](https://huggingface.co/KoReason/KO-REASon-7B-Q2_5-0831) | [Qwen-2.5 (7B)](https://huggingface.co/Qwen/Qwen2.5-7B-Instruct) | `Q2_5` → Qwen-2.5 |
37
 
 
70
  </thead>
71
  <tbody>
72
  <tr>
73
+ <th>KO-REASon-AX3_1-7B-0831(KONI-7B-R-20250831; Ours)</th>
74
  <td>260k</td>
75
  <td>SFT</td>
76
  <td><b>44.6</b></td>
 
85
  <td>38.75</td>
86
  <td><u>49.95</u></td>
87
  </tr>
88
+ <tr>
89
+ <th>KO-REAson-KL3_1-8B-0831(KONI-Llama3.1-8B-R-20250831)</th>
90
+ <td>260k</td>
91
+ <td>SFT</td>
92
+ <td>40.13</td>
93
+ <td>30.57</td>
94
+ <td>43.66</td>
95
+ </tr>
96
  <tr>
97
  <td colspan="6" style="text-align:center; font-weight:bold;">Open Recipe (En)</td>
98
  </tr>
 
170
  <th>AIME 2024</th>
171
  <th>AIME 2025</th>
172
  <th>GPQA</th>
173
+ <th>CLIcK</th>
174
  <th>KoBALT-700</th>
175
  </tr>
176
  </thead>
 
208
  <td>7.61</td><td>28.46</td><td>19.31</td><td>51.61</td><td>46.67</td><td>33.33</td><td>28.28</td><td>32.48</td><td>4.71</td><td>30.61</td><td>27.31</td>
209
  </tr>
210
  <tr>
211
+ <td><b>KO-REAson-KL3_1-8B-0831(KONI-Llama3.1-8B-R-20250831)</b></td>
212
+ <td>8.03</td><td>44.64</td><td>40.08</td><td>37.96</td><td>23.33</td><td>30.00</td><td>38.38</td><td>56.39</td><td>21.57</td><td>30.57</td><td>40.13</td>
213
+ </tr>
214
+ <tr>
215
+ <td><b>KO-REASon-AX3_1-7B-0831 (KONI-7B-R-20250831)</b></td>
216
  <td>7.26</td><td>45.57</td><td>38.13</td><td>52.80</td><td>53.33</td><td>33.33</td><td>36.87</td><td><b>62.86</b></td><td>23.43</td><td><u>43.29</u></td><td><u>44.56</u></td>
217
  </tr>
218
  <tr>