chauhoang commited on
Commit
fc9798b
·
verified ·
1 Parent(s): 261cabb

End of training

Browse files
README.md CHANGED
@@ -64,7 +64,7 @@ lora_model_dir: null
64
  lora_r: 8
65
  lora_target_linear: true
66
  lr_scheduler: cosine
67
- max_steps: 50
68
  micro_batch_size: 2
69
  mlflow_experiment_name: /tmp/b86be1d045906067_train_data.json
70
  model_type: AutoModelForCausalLM
@@ -89,7 +89,7 @@ wandb_name: 94a6cad8-c2cc-425e-aa67-ad52a965d356
89
  wandb_project: Gradients-On-Demand
90
  wandb_run: your_name
91
  wandb_runid: 94a6cad8-c2cc-425e-aa67-ad52a965d356
92
- warmup_steps: 10
93
  weight_decay: 0.0
94
  xformers_attention: null
95
 
@@ -100,8 +100,6 @@ xformers_attention: null
100
  # 94a6cad8-c2cc-425e-aa67-ad52a965d356
101
 
102
  This model is a fine-tuned version of [jhflow/mistral7b-lora-multi-turn-v2](https://huggingface.co/jhflow/mistral7b-lora-multi-turn-v2) on the None dataset.
103
- It achieves the following results on the evaluation set:
104
- - Loss: 0.6012
105
 
106
  ## Model description
107
 
@@ -128,19 +126,14 @@ The following hyperparameters were used during training:
128
  - total_train_batch_size: 8
129
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
130
  - lr_scheduler_type: cosine
131
- - lr_scheduler_warmup_steps: 10
132
- - training_steps: 50
133
 
134
  ### Training results
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
  | No log | 0.0039 | 1 | 0.8528 |
139
- | 3.1245 | 0.0390 | 10 | 0.6828 |
140
- | 2.4105 | 0.0780 | 20 | 0.6242 |
141
- | 2.5341 | 0.1170 | 30 | 0.6087 |
142
- | 2.3267 | 0.1559 | 40 | 0.6020 |
143
- | 2.543 | 0.1949 | 50 | 0.6012 |
144
 
145
 
146
  ### Framework versions
 
64
  lora_r: 8
65
  lora_target_linear: true
66
  lr_scheduler: cosine
67
+ max_steps: 1
68
  micro_batch_size: 2
69
  mlflow_experiment_name: /tmp/b86be1d045906067_train_data.json
70
  model_type: AutoModelForCausalLM
 
89
  wandb_project: Gradients-On-Demand
90
  wandb_run: your_name
91
  wandb_runid: 94a6cad8-c2cc-425e-aa67-ad52a965d356
92
+ warmup_steps: 1
93
  weight_decay: 0.0
94
  xformers_attention: null
95
 
 
100
  # 94a6cad8-c2cc-425e-aa67-ad52a965d356
101
 
102
  This model is a fine-tuned version of [jhflow/mistral7b-lora-multi-turn-v2](https://huggingface.co/jhflow/mistral7b-lora-multi-turn-v2) on the None dataset.
 
 
103
 
104
  ## Model description
105
 
 
126
  - total_train_batch_size: 8
127
  - optimizer: Use OptimizerNames.ADAMW_BNB with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
128
  - lr_scheduler_type: cosine
129
+ - lr_scheduler_warmup_steps: 2
130
+ - training_steps: 1
131
 
132
  ### Training results
133
 
134
  | Training Loss | Epoch | Step | Validation Loss |
135
  |:-------------:|:------:|:----:|:---------------:|
136
  | No log | 0.0039 | 1 | 0.8528 |
 
 
 
 
 
137
 
138
 
139
  ### Framework versions
adapter_config.json CHANGED
@@ -20,13 +20,13 @@
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
23
- "o_proj",
24
  "up_proj",
25
- "gate_proj",
26
  "k_proj",
 
27
  "down_proj",
28
- "v_proj",
29
- "q_proj"
 
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
 
20
  "rank_pattern": {},
21
  "revision": null,
22
  "target_modules": [
 
23
  "up_proj",
 
24
  "k_proj",
25
+ "gate_proj",
26
  "down_proj",
27
+ "q_proj",
28
+ "o_proj",
29
+ "v_proj"
30
  ],
31
  "task_type": "CAUSAL_LM",
32
  "use_dora": false,
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:84dfe489ca26c0721704592b90a0ae07474ff8d09daa76de681cf172b5def105
3
  size 84047370
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31e9e4f307563c755eed62a8071792676961f5dffe12c442a98952b6394bc89a
3
  size 84047370
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:292eadaa61de5dd99e6bf9ec43a5dde3d7c676fce4bff127477199d42e858834
3
  size 83945296
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ef85cd79f7d6a1c7af7c3b5d2da59f4d707d0c6fd2f556b38c62de65f8ba9cf
3
  size 83945296
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68f2d8d172f297e9444c68fa0d9a882ea3108657ab700be8fcd9ce3587988d09
3
  size 6776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ede6b6fdee903a3074da3a171c93ab5343e1258adcee4206296f452294c79e95
3
  size 6776