Uploaded model

  • Developed by: Onuii

  • License: apache-2.0

  • Finetuned from model : kakaocorp/kanana-nano-2.1b-base

    --run_name "DAMI-base-merge-0408"
    --model_name "Onuii/DAMI-base-merge-0408"
    --dataset_name "Onuii/DAMI-Pretrain-Dataset-Tokenize8192"
    --max_seq_length 8192
    --dtype "bfloat16"
    --load_in_4bit True
    --r 128
    --lora_alpha 32
    --lora_dropout 0
    --use_rslora True
    --batch_size 32
    --gradient_accumulation_steps 4
    --learning_rate 1e-4
    --embedding_learning_rate 1e-5
    --warmup_ratio 0.1
    --num_train_epochs 3
    --optimizer "adamw_8bit"
    --weight_decay 0.01
    --lr_scheduler_type "linear"
    --output_dir "../outputs"
    --logging_dir "../logs"
    --logging_steps 1
    --save_steps 200
    --save_total_limit 2
    --report_to "wandb"
    --use_gradient_checkpointing False
    --save_strategy "steps"
    --random_seed 562

This llama model was trained 2x faster with Unsloth and Huggingface's TRL library.

Downloads last month
5
Safetensors
Model size
2.32B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for Onuii/DAMI-base-merge-0408

Finetuned
(6)
this model
Finetunes
2 models