ikerion commited on
Commit
776ae8e
·
verified ·
1 Parent(s): d88c643

Training in progress, step 500

Browse files
adapter_config.json CHANGED
@@ -3,37 +3,29 @@
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
5
  "bias": "none",
6
- "corda_config": null,
7
- "eva_config": null,
8
- "exclude_modules": null,
9
  "fan_in_fan_out": false,
10
  "inference_mode": true,
11
  "init_lora_weights": true,
12
- "layer_replication": null,
13
  "layers_pattern": null,
14
  "layers_to_transform": null,
15
  "loftq_config": {},
16
- "lora_alpha": 64,
17
- "lora_bias": false,
18
  "lora_dropout": 0.05,
19
  "megatron_config": null,
20
  "megatron_core": "megatron.core",
21
  "modules_to_save": null,
22
  "peft_type": "LORA",
23
- "r": 32,
24
  "rank_pattern": {},
25
  "revision": null,
26
  "target_modules": [
27
- "q_proj",
28
- "k_proj",
29
- "down_proj",
30
  "v_proj",
31
- "up_proj",
 
32
  "o_proj",
33
- "gate_proj"
 
 
34
  ],
35
- "task_type": "CAUSAL_LM",
36
- "trainable_token_indices": null,
37
- "use_dora": false,
38
- "use_rslora": false
39
  }
 
3
  "auto_mapping": null,
4
  "base_model_name_or_path": "mistralai/Mistral-7B-Instruct-v0.2",
5
  "bias": "none",
 
 
 
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
8
  "init_lora_weights": true,
 
9
  "layers_pattern": null,
10
  "layers_to_transform": null,
11
  "loftq_config": {},
12
+ "lora_alpha": 128,
 
13
  "lora_dropout": 0.05,
14
  "megatron_config": null,
15
  "megatron_core": "megatron.core",
16
  "modules_to_save": null,
17
  "peft_type": "LORA",
18
+ "r": 64,
19
  "rank_pattern": {},
20
  "revision": null,
21
  "target_modules": [
 
 
 
22
  "v_proj",
23
+ "gate_proj",
24
+ "down_proj",
25
  "o_proj",
26
+ "k_proj",
27
+ "q_proj",
28
+ "up_proj"
29
  ],
30
+ "task_type": "CAUSAL_LM"
 
 
 
31
  }
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0f41f87cf74c110e9e1193011267a840e015a98ec11aaa41114f03de36218605
3
- size 335604696
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0125633087172414d3bfe773c60ec2f28414659891b9a2443eabaaf840f781b5
3
+ size 335605144
runs/Jun14_14-08-15_nio2xbg6s6/events.out.tfevents.1749910096.nio2xbg6s6.292.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:85c58fd4a9104c73114d77ab2d45c5b07208f2d792b69f410d998c118f95df50
3
+ size 4773
tokenizer_config.json CHANGED
@@ -30,13 +30,12 @@
30
  },
31
  "additional_special_tokens": [],
32
  "bos_token": "<s>",
 
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
- "extra_special_tokens": {},
36
  "legacy": false,
37
  "model_max_length": 1000000000000000019884624838656,
38
  "pad_token": "</s>",
39
- "padding_side": "right",
40
  "sp_model_kwargs": {},
41
  "spaces_between_special_tokens": false,
42
  "tokenizer_class": "LlamaTokenizer",
 
30
  },
31
  "additional_special_tokens": [],
32
  "bos_token": "<s>",
33
+ "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{{- bos_token }}\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {%- if loop.first and system_message is defined %}\n {{- ' [INST] ' + system_message + '\\n\\n' + message['content'] + ' [/INST]' }}\n {%- else %}\n {{- ' [INST] ' + message['content'] + ' [/INST]' }}\n {%- endif %}\n {%- elif message['role'] == 'assistant' %}\n {{- ' ' + message['content'] + eos_token}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}\n",
34
  "clean_up_tokenization_spaces": false,
35
  "eos_token": "</s>",
 
36
  "legacy": false,
37
  "model_max_length": 1000000000000000019884624838656,
38
  "pad_token": "</s>",
 
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
41
  "tokenizer_class": "LlamaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3b1f8acc6fee1da9104e6417d454e3c4ab03f3899c67dd6b65d8c131cc31909e
3
- size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:756d407ae9c6d98508652ed40b389920d9a93b3751c9f3b448070a24f0e90620
3
+ size 5048