acezxn commited on
Commit
185728d
·
verified ·
1 Parent(s): 80ac2a5

End of training

Browse files
README.md CHANGED
@@ -4,8 +4,8 @@ library_name: transformers
4
  model_name: SOC_Query_Generation_Base_Llama_3B
5
  tags:
6
  - generated_from_trainer
7
- - unsloth
8
  - trl
 
9
  - sft
10
  licence: license
11
  ---
@@ -35,11 +35,11 @@ This model was trained with SFT.
35
 
36
  ### Framework versions
37
 
38
- - TRL: 0.21.0
39
- - Transformers: 4.55.4
40
  - Pytorch: 2.7.0
41
  - Datasets: 3.6.0
42
- - Tokenizers: 0.21.4
43
 
44
  ## Citations
45
 
 
4
  model_name: SOC_Query_Generation_Base_Llama_3B
5
  tags:
6
  - generated_from_trainer
 
7
  - trl
8
+ - unsloth
9
  - sft
10
  licence: license
11
  ---
 
35
 
36
  ### Framework versions
37
 
38
+ - TRL: 0.23.0
39
+ - Transformers: 4.56.2
40
  - Pytorch: 2.7.0
41
  - Datasets: 3.6.0
42
+ - Tokenizers: 0.22.1
43
 
44
  ## Citations
45
 
adapter_config.json CHANGED
@@ -1,6 +1,10 @@
1
  {
2
  "alpha_pattern": {},
3
- "auto_mapping": null,
 
 
 
 
4
  "base_model_name_or_path": "acezxn/ACI_Cyber_Base_Llama_3B",
5
  "bias": "none",
6
  "corda_config": null,
@@ -25,13 +29,13 @@
25
  "rank_pattern": {},
26
  "revision": null,
27
  "target_modules": [
 
 
 
28
  "o_proj",
29
- "up_proj",
30
  "k_proj",
31
- "down_proj",
32
  "v_proj",
33
- "gate_proj",
34
- "q_proj"
35
  ],
36
  "target_parameters": null,
37
  "task_type": "CAUSAL_LM",
 
1
  {
2
  "alpha_pattern": {},
3
+ "auto_mapping": {
4
+ "base_model_class": "LlamaForCausalLM",
5
+ "parent_library": "transformers.models.llama.modeling_llama",
6
+ "unsloth_fixed": true
7
+ },
8
  "base_model_name_or_path": "acezxn/ACI_Cyber_Base_Llama_3B",
9
  "bias": "none",
10
  "corda_config": null,
 
29
  "rank_pattern": {},
30
  "revision": null,
31
  "target_modules": [
32
+ "q_proj",
33
+ "gate_proj",
34
+ "down_proj",
35
  "o_proj",
 
36
  "k_proj",
 
37
  "v_proj",
38
+ "up_proj"
 
39
  ],
40
  "target_parameters": null,
41
  "task_type": "CAUSAL_LM",
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81509db19f90e91387968101680116fe2655a94547fbdde998b0fec198e6e938
3
  size 194563400
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d95a1c5d6803cff0378d7bf10da1ccf3b5e403f855d9432cf547fcec69c06970
3
  size 194563400
runs/Oct11_17-52-43_19c6da89d26e/events.out.tfevents.1760205179.19c6da89d26e.62545.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e0cf9d5d8768b2e8b04fa0a224f646e61d7fa4495ca800117febbe8c067c9dd
3
+ size 6219
runs/Oct11_17-57-18_19c6da89d26e/events.out.tfevents.1760205441.19c6da89d26e.63683.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:533699c83ff2794a26f9d07b48a71ef210cbf704929558dd979fd7234e62eb40
3
+ size 16919
tokenizer_config.json CHANGED
@@ -2060,7 +2060,7 @@
2060
  ],
2061
  "model_max_length": 131072,
2062
  "pad_token": "<|finetune_right_pad_id|>",
2063
- "padding_side": "right",
2064
  "tokenizer_class": "PreTrainedTokenizerFast",
2065
  "unk_token": null
2066
  }
 
2060
  ],
2061
  "model_max_length": 131072,
2062
  "pad_token": "<|finetune_right_pad_id|>",
2063
+ "padding_side": "left",
2064
  "tokenizer_class": "PreTrainedTokenizerFast",
2065
  "unk_token": null
2066
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:34edf1c5ae1dc1bfbfc0104bbc48fac3dd36edc53b6c73f2aad7cca65d50cc44
3
- size 6289
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69fc783d902c0b19eabcfe518500da7711ea508bf2cbbddaf31e92652c67373d
3
+ size 6353