Training in progress, step 2500
Browse files
adapter_config.json
CHANGED
@@ -25,13 +25,13 @@
|
|
25 |
"revision": null,
|
26 |
"target_modules": [
|
27 |
"query_key_value",
|
|
|
|
|
28 |
"o_proj",
|
29 |
"gate_proj",
|
30 |
-
"up_proj",
|
31 |
-
"k_proj",
|
32 |
"q_proj",
|
33 |
-
"
|
34 |
-
"
|
35 |
],
|
36 |
"task_type": "CAUSAL_LM",
|
37 |
"trainable_token_indices": null,
|
|
|
25 |
"revision": null,
|
26 |
"target_modules": [
|
27 |
"query_key_value",
|
28 |
+
"down_proj",
|
29 |
+
"k_proj",
|
30 |
"o_proj",
|
31 |
"gate_proj",
|
|
|
|
|
32 |
"q_proj",
|
33 |
+
"v_proj",
|
34 |
+
"up_proj"
|
35 |
],
|
36 |
"task_type": "CAUSAL_LM",
|
37 |
"trainable_token_indices": null,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 360740440
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:47feebcdb604aa8140889d5be160c347f07305566fd0a219036cb87ac58d58b0
|
3 |
size 360740440
|
toilaluan/7a6f2833-62b5-4c10-813b-08b22ac32093/adapter_config.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"alpha_pattern": {}, "auto_mapping": null, "base_model_name_or_path": "unsloth/Llama-3.2-1B-Instruct", "bias": "none", "fan_in_fan_out": false, "inference_mode": true, "init_lora_weights": true, "layer_replication": null, "layers_pattern": null, "layers_to_transform": null, "loftq_config": {}, "lora_alpha": 256, "lora_dropout": 0.0, "megatron_config": null, "megatron_core": "megatron.core", "modules_to_save": null, "peft_type": "LORA", "r": 128, "rank_pattern": {}, "revision": null, "target_modules": ["
|
|
|
1 |
+
{"alpha_pattern": {}, "auto_mapping": null, "base_model_name_or_path": "unsloth/Llama-3.2-1B-Instruct", "bias": "none", "fan_in_fan_out": false, "inference_mode": true, "init_lora_weights": true, "layer_replication": null, "layers_pattern": null, "layers_to_transform": null, "loftq_config": {}, "lora_alpha": 256, "lora_dropout": 0.0, "megatron_config": null, "megatron_core": "megatron.core", "modules_to_save": null, "peft_type": "LORA", "r": 128, "rank_pattern": {}, "revision": null, "target_modules": ["v_proj", "q_proj", "gate_proj", "o_proj", "k_proj", "up_proj", "down_proj", "query_key_value"], "task_type": "CAUSAL_LM", "use_dora": false, "use_rslora": false}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8d20c4bf9877d35554600ab548bd0f527c55f2e8f71f8ee163111ea7ebb3b4a
|
3 |
+
size 5816
|