|
--- |
|
model_creator: Nekochu |
|
quantized_by: Nekochu |
|
model_name: Llama-2 13B German ORPO |
|
pretty_name: Llama-2 13B German ORPO |
|
model_type: llama2 |
|
prompt_template: >- |
|
Below is an instruction that describes a task. Write a response that |
|
appropriately completes the request. ### Instruction: {Instruction} {summary} ### input: {category} ### Response: {prompt} |
|
base_model: meta-llama/Llama-2-13b-chat-hf |
|
library_name: peft |
|
license: apache-2.0 |
|
datasets: |
|
- mayflowergmbh/intel_orca_dpo_pairs_de |
|
- LeoLM/OpenSchnabeltier |
|
- LeoLM/German_Songs |
|
- LeoLM/German_Poems |
|
- bjoernp/ultrachat_de |
|
- mayflowergmbh/ultra-chat_de |
|
- mayflowergmbh/airoboros-3.0_de |
|
- mayflowergmbh/booksum_de |
|
- mayflowergmbh/dolphin_de |
|
- mayflowergmbh/evol-instruct_de |
|
- mayflowergmbh/openschnabeltier_de |
|
- mayflowergmbh/alpaca-gpt4_de |
|
- mayflowergmbh/dolly-15k_de |
|
- mayflowergmbh/oasst_de |
|
language: |
|
- de |
|
- en |
|
pipeline_tag: text-generation |
|
task_categories: |
|
- question-answering |
|
- text2text-generation |
|
- conversational |
|
inference: True |
|
tags: |
|
- llama-factory |
|
- lora |
|
- generated_from_trainer |
|
- llama2 |
|
- llama |
|
- instruct |
|
- finetune |
|
- llm |
|
- pytorch |
|
- llama |
|
- llama-2 |
|
- german |
|
- deutsch |
|
model-index: |
|
- name: Llama-2-13B-German-ORPO |
|
results: [] |
|
--- |
|
|
|
llama2 Chat LoRa sft train Stage A on German dataset: |
|
```text |
|
German_Songs,German_Poems,bjoernp_ultrachat_de,OpenSchnabeltier,ultrachat_de,oasst_de,dolly_15k_de,alpaca-gpt4_de,openschnabeltier_de,evol_instruct_de,dolphin_de,booksum_de,airoboros_de & eval VAGOsolutions/MT-Bench-TrueGerman? |
|
``` |
|
|
|
Stage B: Resume LoRa training using ORPO and dataset `mayflowergmbh/intel_orca_dpo_pairs_de` |
|
|
|
Oh and I am not GER speaker ^^ |
|
|
|
### Training hyperparameters |
|
|
|
```cmd |
|
python src/train_bash.py --stage sft ... --finetuning_type lora --quantization_bit 4 --template alpaca --rope_scaling linear --flash_attn True --dataset_dir data --dataset German_Songs,German_Poems,bjoernp_ultrachat_de,OpenSchnabeltier,ultrachat_de,oasst_de,dolly_15k_de,alpaca-gpt4_de,openschnabeltier_de,evol_instruct_de,dolphin_de,booksum_de,airoboros_de --cutoff_len 4096 --learning_rate 5e-05 --num_train_epochs 1.0 --max_samples 100000 --per_device_train_batch_size 1 --gradient_accumulation_steps 1 --lr_scheduler_type cosine --max_grad_norm 1.0 --logging_steps 5 --save_steps 1000 --warmup_steps 0 --neftune_noise_alpha 0.5 --optim adamw_torch --upcast_layernorm True --use_llama_pro True --bf16 True --lora_rank 512 --lora_alpha 1024 --lora_dropout 0.15 --lora_target all --use_rslora True --additional_target all --create_new_adapter True --plot_loss True |
|
|
|
python src/train_bash.py --stage orpo ... --finetuning_type lora --quantization_bit 4 --template alpaca --rope_scaling linear --flash_attn True --dataset_dir data --dataset orca_dpo_de --cutoff_len 4096 --learning_rate 1e-05 --num_train_epochs 1.0 --max_samples 100000 --per_device_train_batch_size 1 --gradient_accumulation_steps 1 --lr_scheduler_type cosine --max_grad_norm 0.9 --logging_steps 5 --save_steps 250 --warmup_steps 100 --neftune_noise_alpha 0.5 --optim adamw_torch --upcast_layernorm True --use_llama_pro True --report_to none --bf16 True --lora_rank 512 --lora_alpha 1024 --lora_dropout 0.15 --use_rslora True --lora_target all --additional_target all --orpo_beta 0.1 --plot_loss True |
|
``` |
|
The following hyperparameters were used during training: |
|
- learning_rate: 1e-05 # not Defaut LR as for high rank 512, alpha 1024 |
|
- train_batch_size: 1 |
|
- eval_batch_size: 8 |
|
- seed: 42 |
|
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08 |
|
- lr_scheduler_type: cosine |
|
- lr_scheduler_warmup_steps: 1000 |
|
- num_epochs: 1.0 |
|
### Framework versions |
|
- PEFT 0.10.0 |
|
- Transformers 4.39.1 |
|
- Pytorch 2.2.1+cu121 |
|
- Datasets 2.18.0 |
|
- Tokenizers 0.15.2 |