A newer version of this model is available:
unsloth/gemma-2-2b-it-bnb-4bit
Gorani Model Card
์๊ฐ (Introduce)
์ด ๋ชจ๋ธ์ ๋ฒ์ญ์ ์ํ ๋ชจ๋ธ์
๋๋ค. ํ๊ตญ ๊ณ ์ ์ด์ ์ ํํ ๋ฒ์ญ์ ์์ฑํ๊ธฐ ์ํด ํ๊ตญ์ด, ์์ด, ์ผ๋ณธ์ด์ ์ธ์ด ๋ฐ์ดํฐ๋ฅผ ํผํฉํ์ฌ unsloth/gemma-2b-it-bnb-4bit์ ํ์ต์์ผ ์์ฑ๋ gorani-2B ์
๋๋ค.
gorani๋ ํ์ฌ ํ๊ตญ์ด, ์์ด, ์ผ๋ณธ์ด๋ง ๋ฒ์ญ์ ์ง์ํฉ๋๋ค.
๋ชจ๋ธ ์ ๋ณด
- ๊ฐ๋ฐ์: taeyoon0620
- ๋ชจ๋ธ ์ ํ: gemma๋ฅผ ๊ธฐ๋ฐ์ผ๋ก ํ๋ 2B ๋งค๊ฐ๋ณ์ ๋ชจ๋ธ์ธ gorani-2B
- ์ง์ ์ธ์ด: ํ๊ตญ์ด, ์์ด, ์ผ๋ณธ์ด
- ๋ผ์ด์ผ์ค: gemma
Training Hyperparameters
- per_device_train_batch_size: 8
- gradient_accumulation_steps: 4
- warmup_steps: 5
- learning_rate: 2e-5
- fp16:
not is_bfloat16_supported()
- num_train_epochs: 3
- weight_decay: 0.01
- lr_scheduler_type: "linear"
ํ์ต ๋ฐ์ดํฐ
ํ์ต ์ฑ๋ฅ ๋น๊ต
Step | Training Loss | Validation Loss |
---|---|---|
50 | 0.856900 | 1.243296 |
100 | 0.637300 | 1.080636 |
150 | 0.561300 | 1.015324 |
200 | 0.523100 | 0.972539 |
250 | 0.491600 | 0.951547 |
300 | 0.485000 | 0.942520 |
350 | 0.457100 | 0.928466 |
400 | 0.482600 | 0.913566 |
450 | 0.445000 | 0.915080 |
500 | 0.447300 | 0.900509 |
550 | 0.443200 | 0.896823 |
600 | 0.420800 | 0.898253 |
650 | 0.430200 | 0.888035 |
700 | 0.428700 | 0.884415 |
750 | 0.416400 | 0.875832 |
800 | 0.397400 | 0.877212 |
850 | 0.403800 | 0.878696 |
900 | 0.410400 | 0.873603 |
950 | 0.423900 | 0.876391 |
1000 | 0.412100 | 0.874712 |
1050 | 0.404000 | 0.878219 |
1100 | 0.391000 | 0.868475 |
1150 | 0.383800 | 0.867502 |
1200 | 0.388800 | 0.877054 |
1250 | 0.382200 | 0.864894 |
1300 | 0.389500 | 0.862716 |
1350 | 0.401100 | 0.857156 |
1400 | 0.364600 | 0.865956 |
1450 | 0.386700 | 0.858560 |
1500 | 0.379200 | 0.860608 |
1550 | 0.370000 | 0.864580 |
1600 | 0.367900 | 0.863778 |
1650 | 0.360100 | 0.866951 |
1700 | 0.372300 | 0.867361 |
1750 | 0.378400 | 0.867690 |
1800 | 0.371100 | 0.863734 |
1850 | 0.376400 | 0.862673 |
1900 | 0.364200 | 0.862987 |
1950 | 0.368000 | 0.865702 |
2000 | 0.374900 | 0.865652 |
- Downloads last month
- 0
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
๐
Ask for provider support
HF Inference deployability: The model has no pipeline_tag.
Model tree for taeyoon12321421/gorani-gemma-model-2b
Base model
unsloth/gemma-2b-it-bnb-4bit