kalomaze commited on
Commit
27847b3
·
verified ·
1 Parent(s): 7b871f3

Upload rm_v2.yaml with huggingface_hub

Browse files
Files changed (1) hide show
  1. rm_v2.yaml +68 -0
rm_v2.yaml ADDED
@@ -0,0 +1,68 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ base_model: /ephemeral/qwen_base/gate_reduction_middle_outward_init
2
+ base_model_ignore_patterns: "*/*"
3
+ # optionally might have model_type or tokenizer_type
4
+ model_type: AutoModelForCausalLM
5
+ tokenizer_type: AutoTokenizer
6
+ trust_remote_code: true
7
+
8
+ plugins:
9
+ - axolotl.integrations.liger.LigerPlugin
10
+
11
+ liger_rope: true
12
+ liger_rms_norm: true
13
+ liger_swiglu: true
14
+ liger_fused_linear_cross_entropy: false
15
+
16
+ load_in_8bit: false
17
+ load_in_4bit: false
18
+ strict: false
19
+
20
+ auto_resume_from_checkpoints: true
21
+
22
+ datasets:
23
+ - path: /ephemeral/boom/verifiers/quest-tools/data_rm_tools
24
+ data_files: "/ephemeral/boom/verifiers/quest-tools/data_rm_tools/sample_heha_20k_axo.jsonl"
25
+ ds_type: json
26
+ type: completion
27
+ use_hub: false
28
+
29
+ shuffle_merged_datasets: true
30
+ dataset_prepared_path: ./prepared_v2-12/
31
+ val_set_size: 0.005
32
+ output_dir: ./output-20k-both-2/
33
+ sequence_len: 6144
34
+ sample_packing: true
35
+ eval_sample_packing: true
36
+
37
+ gradient_accumulation_steps: 1
38
+ micro_batch_size: 4
39
+ num_epochs: 2
40
+ optimizer: paged_adamw_8bit
41
+ lr_scheduler: cosine
42
+
43
+ learning_rate: 1e-6
44
+ adam_beta1: 0.9
45
+ adam_beta2: 0.999
46
+ max_grad_norm: 0.001
47
+ train_on_inputs: false
48
+ group_by_length: false
49
+ bf16: auto
50
+
51
+ gradient_checkpointing: true
52
+ gradient_checkpointing_kwargs:
53
+ use_reentrant: true
54
+
55
+ logging_steps: 1
56
+ flash_attention: true
57
+ warmup_steps: 200
58
+ eval_steps: 50
59
+ saves_per_epoch: 5
60
+
61
+ deepspeed: /ephemeral/axolotl/axolotl/deepspeed_configs/zero3_bf16.json
62
+
63
+ auto_resume_from_checkpoints: false
64
+ wandb_project: corruption_model_rm_chunk
65
+ wandb_entity:
66
+ wandb_watch:
67
+ wandb_name: rm-v2-20k-both-2
68
+ wandb_log_model: