mhenrichsen commited on
Commit
103b340
·
verified ·
1 Parent(s): 732d740

End of training

Browse files
Files changed (1) hide show
  1. README.md +4 -4
README.md CHANGED
@@ -60,7 +60,7 @@ resume_from_checkpoint:
60
  logging_steps: 1
61
  flash_attention: true
62
 
63
- warmup_steps: 10
64
  evals_per_epoch: 2
65
  saves_per_epoch: 1
66
  weight_decay: 0.0
@@ -75,7 +75,7 @@ special_tokens:
75
 
76
  This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the syvai/no-emotion-reasoning dataset.
77
  It achieves the following results on the evaluation set:
78
- - Loss: 1.9291
79
 
80
  ## Model description
81
 
@@ -102,7 +102,7 @@ The following hyperparameters were used during training:
102
  - total_train_batch_size: 16
103
  - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
104
  - lr_scheduler_type: cosine
105
- - lr_scheduler_warmup_steps: 10
106
  - num_epochs: 1.0
107
 
108
  ### Training results
@@ -110,7 +110,7 @@ The following hyperparameters were used during training:
110
  | Training Loss | Epoch | Step | Validation Loss |
111
  |:-------------:|:------:|:----:|:---------------:|
112
  | 8.4068 | 0.1013 | 1 | 8.7623 |
113
- | 3.8505 | 0.5063 | 5 | 1.9291 |
114
 
115
 
116
  ### Framework versions
 
60
  logging_steps: 1
61
  flash_attention: true
62
 
63
+ warmup_steps: 1
64
  evals_per_epoch: 2
65
  saves_per_epoch: 1
66
  weight_decay: 0.0
 
75
 
76
  This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the syvai/no-emotion-reasoning dataset.
77
  It achieves the following results on the evaluation set:
78
+ - Loss: 0.5303
79
 
80
  ## Model description
81
 
 
102
  - total_train_batch_size: 16
103
  - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
104
  - lr_scheduler_type: cosine
105
+ - lr_scheduler_warmup_steps: 2
106
  - num_epochs: 1.0
107
 
108
  ### Training results
 
110
  | Training Loss | Epoch | Step | Validation Loss |
111
  |:-------------:|:------:|:----:|:---------------:|
112
  | 8.4068 | 0.1013 | 1 | 8.7623 |
113
+ | 0.8452 | 0.5063 | 5 | 0.5303 |
114
 
115
 
116
  ### Framework versions