A newer version of this model is available: unsloth/Llama-3.1-8B-Instruct-unsloth-bnb-4bit

Gorani Model Card

์†Œ๊ฐœ (Introduce)

์ด ๋ชจ๋ธ์€ ๋ฒˆ์—ญ์„ ์œ„ํ•œ ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค. ํ•œ๊ตญ ๊ณ ์œ ์–ด์˜ ์ •ํ™•ํ•œ ๋ฒˆ์—ญ์„ ์ƒ์„ฑํ•˜๊ธฐ ์œ„ํ•ด ํ•œ๊ตญ์–ด, ์˜์–ด, ์ผ๋ณธ์–ด์˜ ์–ธ์–ด ๋ฐ์ดํ„ฐ๋ฅผ ํ˜ผํ•ฉํ•˜์—ฌ unsloth/Llama-3.1-8B-Instruct-unsloth-bnb-4bit์„ ํ•™์Šต์‹œ์ผœ ์ƒ์„ฑ๋œ gorani-8b-new ์ž…๋‹ˆ๋‹ค.
gorani๋Š” ํ˜„์žฌ ํ•œ๊ตญ์–ด, ์˜์–ด, ์ผ๋ณธ์–ด๋งŒ ๋ฒˆ์—ญ์„ ์ง€์›ํ•ฉ๋‹ˆ๋‹ค.

๋ชจ๋ธ ์ •๋ณด

  • ๊ฐœ๋ฐœ์ž: taeyoon0620
  • ๋ชจ๋ธ ์œ ํ˜•: llama๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•˜๋Š” 8B ๋งค๊ฐœ๋ณ€์ˆ˜ ๋ชจ๋ธ์ธ gorani-8B
  • ์ง€์› ์–ธ์–ด: ํ•œ๊ตญ์–ด, ์˜์–ด, ์ผ๋ณธ์–ด
  • ๋ผ์ด์„ผ์Šค: llama

Training Hyperparameters

  • per_device_train_batch_size: 8
  • gradient_accumulation_steps: 4
  • warmup_steps: 5
  • learning_rate: 2e-5
  • fp16: not is_bfloat16_supported()
  • num_train_epochs: 3
  • weight_decay: 0.01
  • lr_scheduler_type: "linear"

image/png

ํ•™์Šต ๋ฐ์ดํ„ฐ

๋ฐ์ดํ„ฐ์…‹ ๋งํฌ

ํ•™์Šต ์„ฑ๋Šฅ ๋น„๊ต

Downloads last month
0
Safetensors
Model size
8.03B params
Tensor type
BF16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for taeyoon12321421/gorani-8b-merged-16bit

Dataset used to train taeyoon12321421/gorani-8b-merged-16bit