vpakarinen commited on
Commit
20c0166
·
verified ·
1 Parent(s): e9de600

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -39
README.md CHANGED
@@ -92,42 +92,4 @@ weight_decay: 0.0
92
 
93
  # outputs/mymodel
94
 
95
- This model is a fine-tuned version of [unsloth/Qwen2.5-7B-Instruct](https://huggingface.co/unsloth/Qwen2.5-7B-Instruct) on the vpakarinen/uncensored-tech dataset.
96
-
97
- ## Model description
98
-
99
- More information needed
100
-
101
- ## Intended uses & limitations
102
-
103
- More information needed
104
-
105
- ## Training and evaluation data
106
-
107
- More information needed
108
-
109
- ## Training procedure
110
-
111
- ### Training hyperparameters
112
-
113
- The following hyperparameters were used during training:
114
- - learning_rate: 0.0002
115
- - train_batch_size: 8
116
- - eval_batch_size: 8
117
- - seed: 42
118
- - optimizer: Use OptimizerNames.PAGED_ADAMW_8BIT with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
119
- - lr_scheduler_type: cosine
120
- - lr_scheduler_warmup_steps: 7
121
- - num_epochs: 1.0
122
-
123
- ### Training results
124
-
125
-
126
-
127
- ### Framework versions
128
-
129
- - PEFT 0.14.0
130
- - Transformers 4.49.0
131
- - Pytorch 2.5.1+cu124
132
- - Datasets 3.2.0
133
- - Tokenizers 0.21.0
 
92
 
93
  # outputs/mymodel
94
 
95
+ This model is a fine-tuned version of [unsloth/Qwen2.5-7B-Instruct](https://huggingface.co/unsloth/Qwen2.5-7B-Instruct) on the vpakarinen/uncensored-tech dataset.