NanQiangHF commited on
Commit
f3c9f18
·
verified ·
1 Parent(s): b4553ba

llama3.1_8b_bwgenerator

Browse files
README.md CHANGED
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.1032
22
 
23
  ## Model description
24
 
@@ -51,14 +51,14 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:------:|:----:|:---------------:|
54
- | 0.7124 | 0.1216 | 40 | 0.2698 |
55
- | 0.2279 | 0.2433 | 80 | 0.1875 |
56
- | 0.1546 | 0.3649 | 120 | 0.1293 |
57
- | 0.1242 | 0.4865 | 160 | 0.1168 |
58
- | 0.1144 | 0.6081 | 200 | 0.1103 |
59
- | 0.1097 | 0.7298 | 240 | 0.1064 |
60
- | 0.107 | 0.8514 | 280 | 0.1043 |
61
- | 0.105 | 0.9730 | 320 | 0.1032 |
62
 
63
 
64
  ### Framework versions
 
18
 
19
  This model is a fine-tuned version of [meta-llama/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3.1-8B-Instruct) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.0982
22
 
23
  ## Model description
24
 
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss |
53
  |:-------------:|:------:|:----:|:---------------:|
54
+ | 0.7155 | 0.1216 | 40 | 0.2546 |
55
+ | 0.218 | 0.2433 | 80 | 0.1804 |
56
+ | 0.1513 | 0.3649 | 120 | 0.1246 |
57
+ | 0.1193 | 0.4865 | 160 | 0.1116 |
58
+ | 0.1092 | 0.6081 | 200 | 0.1051 |
59
+ | 0.1046 | 0.7298 | 240 | 0.1012 |
60
+ | 0.1017 | 0.8514 | 280 | 0.0993 |
61
+ | 0.0999 | 0.9730 | 320 | 0.0982 |
62
 
63
 
64
  ### Framework versions
adapter_config.json CHANGED
@@ -22,8 +22,8 @@
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
- "v_proj",
26
- "q_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
 
22
  "rank_pattern": {},
23
  "revision": null,
24
  "target_modules": [
25
+ "q_proj",
26
+ "v_proj"
27
  ],
28
  "task_type": "CAUSAL_LM",
29
  "use_dora": false,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:93de4bf57aab1828667bacd328921a78ab4e52d693f424d7d0ba794e8a949deb
3
  size 6832728
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1de99ba3c69896469e24e31d640496d977ca9154e6e8ede2c9d8c58ee1c49a20
3
  size 6832728
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:94b7f1252de2ef7de0ca1f0f926c6ab823e3b4e41e07becfca600b0eb3228a3e
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d289afc35be50448ca012c8270b76abc4f7753d1d7bd83a50c3267c0533c498d
3
  size 5496