{ "base_model": "unsloth/llama-3-8b-bnb-4bit", "dataset": "metythorn/khmer-corpus", "training_framework": "unsloth", "lora_config": { "r": 128, "lora_alpha": 32, "lora_dropout": 0, "bias": "none", "target_modules": [ "q_proj", "k_proj", "v_proj", "o_proj", "gate_proj", "up_proj", "down_proj", "embed_proj", "lm_head" ], "use_rslora": true }, "training_args": { "per_device_train_batch_size": 2, "gradient_accumulation_steps": 8, "warmup_ratio": 0.1, "num_train_epochs": 1, "learning_rate": 5e-05, "embedding_learning_rate": 5e-06, "optim": "adamw_8bit", "lr_scheduler_type": "cosine", "weight_decay": 0.0, "seed": 3407 }, "model_config": { "max_seq_length": 2048, "load_in_4bit": true, "use_gradient_checkpointing": "unsloth" } }