RayBernard commited on
Commit
ec7ae49
·
verified ·
1 Parent(s): e7a1581

Model save

Browse files
README.md CHANGED
@@ -1,11 +1,10 @@
1
  ---
2
- license: llama3
3
  library_name: peft
4
  tags:
5
  - trl
6
  - sft
7
  - generated_from_trainer
8
- base_model: meta-llama/Meta-Llama-3-8B-Instruct
9
  model-index:
10
  - name: llama-3-8B-Instruct-ft
11
  results: []
@@ -16,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # llama-3-8B-Instruct-ft
18
 
19
- This model is a fine-tuned version of [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) on an unknown dataset.
20
 
21
  ## Model description
22
 
@@ -44,7 +43,7 @@ The following hyperparameters were used during training:
44
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
45
  - lr_scheduler_type: cosine
46
  - lr_scheduler_warmup_ratio: 0.1
47
- - training_steps: 200
48
  - mixed_precision_training: Native AMP
49
 
50
  ### Framework versions
 
1
  ---
 
2
  library_name: peft
3
  tags:
4
  - trl
5
  - sft
6
  - generated_from_trainer
7
+ base_model: final-tuned-model
8
  model-index:
9
  - name: llama-3-8B-Instruct-ft
10
  results: []
 
15
 
16
  # llama-3-8B-Instruct-ft
17
 
18
+ This model was trained from scratch on an unknown dataset.
19
 
20
  ## Model description
21
 
 
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: cosine
45
  - lr_scheduler_warmup_ratio: 0.1
46
+ - training_steps: 150
47
  - mixed_precision_training: Native AMP
48
 
49
  ### Framework versions
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct",
5
  "bias": "all",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
@@ -20,9 +20,9 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "output_proj",
24
  "v_proj",
25
- "q_proj"
 
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "final-tuned-model",
5
  "bias": "all",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "v_proj",
24
+ "q_proj",
25
+ "output_proj"
26
  ],
27
  "task_type": "CAUSAL_LM",
28
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:58524a0607e2a17c97e70d17ac1bea5ed5b5a6314c6a4e9d4d892ab5308f37d1
3
- size 4308387424
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd05f214dc622562ed168c3655734276c2266553db79edc68825ad1448fdaaca
3
+ size 105661296
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc1badfe9164fafe3b2adabb5e7b208234404b8d3ecc77196e9c07a31d93703a
3
  size 5368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ea75da07f1ba70e77c160385fa626b641050435114d80aa7f5462afa04741a30
3
  size 5368