Update README.md
Browse files
README.md
CHANGED
|
@@ -33,11 +33,8 @@ Our goal is to empower anyone to build and advance their own large language mode
|
|
| 33 |
Key Features:
|
| 34 |
|
| 35 |
1. A 10B-parameter Korean–English reasoning model trained entirely from scratch.
|
| 36 |
-
|
| 37 |
2. 100% open resources — including all training data, code, intermediate checkpoints, and tutorials — allowing anyone to reproduce and extend a near-SOTA model on their own.
|
| 38 |
-
|
| 39 |
3. 3 trillion tokens of training data released publicly, featuring never-before-shared, high-quality full-cycle Korean datasets (for pretraining, post-training, general, reasoning, and reinforcement learning).
|
| 40 |
-
|
| 41 |
4. A collaborative effort by eight undergraduate and master’s students at the KAIST Graduate School of Culture Technology (MLP Lab), documented in a 45-page research paper.
|
| 42 |
|
| 43 |
If you’ve ever used a Korean language model that performs well on benchmarks but feels strange in real use, or if fine-tuning only made it worse, you’re not alone.
|
|
@@ -45,7 +42,7 @@ If you’ve ever used a Korean language model that performs well on benchmarks b
|
|
| 45 |
KORMo solves these problems head-on.
|
| 46 |
By releasing every intermediate model and post-training dataset, we give users the freedom to build on the base model with their own data, customizing and fine-tuning it in any direction they want.
|
| 47 |
|
| 48 |
-
👉
|
| 49 |
```
|
| 50 |
|
| 51 |
---
|
|
|
|
| 33 |
Key Features:
|
| 34 |
|
| 35 |
1. A 10B-parameter Korean–English reasoning model trained entirely from scratch.
|
|
|
|
| 36 |
2. 100% open resources — including all training data, code, intermediate checkpoints, and tutorials — allowing anyone to reproduce and extend a near-SOTA model on their own.
|
|
|
|
| 37 |
3. 3 trillion tokens of training data released publicly, featuring never-before-shared, high-quality full-cycle Korean datasets (for pretraining, post-training, general, reasoning, and reinforcement learning).
|
|
|
|
| 38 |
4. A collaborative effort by eight undergraduate and master’s students at the KAIST Graduate School of Culture Technology (MLP Lab), documented in a 45-page research paper.
|
| 39 |
|
| 40 |
If you’ve ever used a Korean language model that performs well on benchmarks but feels strange in real use, or if fine-tuning only made it worse, you’re not alone.
|
|
|
|
| 42 |
KORMo solves these problems head-on.
|
| 43 |
By releasing every intermediate model and post-training dataset, we give users the freedom to build on the base model with their own data, customizing and fine-tuning it in any direction they want.
|
| 44 |
|
| 45 |
+
👉 "If you want a great Korean language model, now you can build it yourself. It even works with free Colab GPUs!" 🤗
|
| 46 |
```
|
| 47 |
|
| 48 |
---
|