mjkmain commited on
Commit
f064525
·
verified ·
1 Parent(s): 26b0cf9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -4
README.md CHANGED
@@ -33,11 +33,8 @@ Our goal is to empower anyone to build and advance their own large language mode
33
  Key Features:
34
 
35
  1. A 10B-parameter Korean–English reasoning model trained entirely from scratch.
36
-
37
  2. 100% open resources — including all training data, code, intermediate checkpoints, and tutorials — allowing anyone to reproduce and extend a near-SOTA model on their own.
38
-
39
  3. 3 trillion tokens of training data released publicly, featuring never-before-shared, high-quality full-cycle Korean datasets (for pretraining, post-training, general, reasoning, and reinforcement learning).
40
-
41
  4. A collaborative effort by eight undergraduate and master’s students at the KAIST Graduate School of Culture Technology (MLP Lab), documented in a 45-page research paper.
42
 
43
  If you’ve ever used a Korean language model that performs well on benchmarks but feels strange in real use, or if fine-tuning only made it worse, you’re not alone.
@@ -45,7 +42,7 @@ If you’ve ever used a Korean language model that performs well on benchmarks b
45
  KORMo solves these problems head-on.
46
  By releasing every intermediate model and post-training dataset, we give users the freedom to build on the base model with their own data, customizing and fine-tuning it in any direction they want.
47
 
48
- 👉 If you want a great Korean language model, now you can build it yourself. It even works with free Colab GPUs!” 🤗
49
  ```
50
 
51
  ---
 
33
  Key Features:
34
 
35
  1. A 10B-parameter Korean–English reasoning model trained entirely from scratch.
 
36
  2. 100% open resources — including all training data, code, intermediate checkpoints, and tutorials — allowing anyone to reproduce and extend a near-SOTA model on their own.
 
37
  3. 3 trillion tokens of training data released publicly, featuring never-before-shared, high-quality full-cycle Korean datasets (for pretraining, post-training, general, reasoning, and reinforcement learning).
 
38
  4. A collaborative effort by eight undergraduate and master’s students at the KAIST Graduate School of Culture Technology (MLP Lab), documented in a 45-page research paper.
39
 
40
  If you’ve ever used a Korean language model that performs well on benchmarks but feels strange in real use, or if fine-tuning only made it worse, you’re not alone.
 
42
  KORMo solves these problems head-on.
43
  By releasing every intermediate model and post-training dataset, we give users the freedom to build on the base model with their own data, customizing and fine-tuning it in any direction they want.
44
 
45
+ 👉 "If you want a great Korean language model, now you can build it yourself. It even works with free Colab GPUs!" 🤗
46
  ```
47
 
48
  ---