nblinh63 commited on
Commit
740958b
·
verified ·
1 Parent(s): b95b75f

End of training

Browse files
README.md CHANGED
@@ -95,7 +95,7 @@ xformers_attention: true
95
 
96
  This model is a fine-tuned version of [jhflow/mistral7b-lora-multi-turn-v2](https://huggingface.co/jhflow/mistral7b-lora-multi-turn-v2) on the None dataset.
97
  It achieves the following results on the evaluation set:
98
- - Loss: 1.8362
99
 
100
  ## Model description
101
 
@@ -127,7 +127,7 @@ The following hyperparameters were used during training:
127
 
128
  | Training Loss | Epoch | Step | Validation Loss |
129
  |:-------------:|:------:|:----:|:---------------:|
130
- | 1.848 | 0.0016 | 100 | 1.8362 |
131
 
132
 
133
  ### Framework versions
 
95
 
96
  This model is a fine-tuned version of [jhflow/mistral7b-lora-multi-turn-v2](https://huggingface.co/jhflow/mistral7b-lora-multi-turn-v2) on the None dataset.
97
  It achieves the following results on the evaluation set:
98
+ - Loss: 1.8317
99
 
100
  ## Model description
101
 
 
127
 
128
  | Training Loss | Epoch | Step | Validation Loss |
129
  |:-------------:|:------:|:----:|:---------------:|
130
+ | 1.8513 | 0.0016 | 100 | 1.8317 |
131
 
132
 
133
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "q_proj",
24
- "gate_proj",
25
- "down_proj",
26
  "up_proj",
 
27
  "o_proj",
28
  "k_proj",
29
- "v_proj"
 
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
 
 
23
  "up_proj",
24
+ "v_proj",
25
  "o_proj",
26
  "k_proj",
27
+ "gate_proj",
28
+ "down_proj",
29
+ "q_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a7ccde65b8c2136901556b0f53f4d4c2e278b6625cdca0ba0d510a3fd73c7117
3
  size 167934026
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7490cd22949dec8082cd7a645fa7e9065a178fe31fab35b12d4bd983f5cd4e6d
3
  size 167934026
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6eb94cde3497c6e05a8c13e571ddcdaf6c618e76b7da9d07d1e0846571c9df41
3
  size 167832240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2dedd79843b57edf1439a9847d9c95cfbf1d26974d31a4552e30777d952c2093
3
  size 167832240
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b5a3017c6a6a9ed7332dc98650889e479ca7a6407699468821088653ca533319
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:77301c09dda99d8849dca644a17acf18db3b149749b2da4513ad8afde91b7f4f
3
  size 6776