Update README.md
Browse files
README.md
CHANGED
|
@@ -49,6 +49,7 @@ From the Open-Platypus dataset filtering-out rows which has leetcode_ne in it's
|
|
| 49 |
#### Training Hyperparameters
|
| 50 |
|
| 51 |
LoraConfig(
|
|
|
|
| 52 |
r=4,
|
| 53 |
lora_alpha=2,
|
| 54 |
target_modules=modules,
|
|
@@ -58,6 +59,7 @@ LoraConfig(
|
|
| 58 |
)
|
| 59 |
|
| 60 |
TrainingArguments(
|
|
|
|
| 61 |
output_dir="gemma-2b-code-platypus",
|
| 62 |
num_train_epochs=1,
|
| 63 |
per_device_train_batch_size=4,
|
|
@@ -78,6 +80,7 @@ TrainingArguments(
|
|
| 78 |
)
|
| 79 |
|
| 80 |
SFTTrainer(
|
|
|
|
| 81 |
model=model,
|
| 82 |
train_dataset=train_data,
|
| 83 |
eval_dataset=test_data,
|
|
|
|
| 49 |
#### Training Hyperparameters
|
| 50 |
|
| 51 |
LoraConfig(
|
| 52 |
+
|
| 53 |
r=4,
|
| 54 |
lora_alpha=2,
|
| 55 |
target_modules=modules,
|
|
|
|
| 59 |
)
|
| 60 |
|
| 61 |
TrainingArguments(
|
| 62 |
+
|
| 63 |
output_dir="gemma-2b-code-platypus",
|
| 64 |
num_train_epochs=1,
|
| 65 |
per_device_train_batch_size=4,
|
|
|
|
| 80 |
)
|
| 81 |
|
| 82 |
SFTTrainer(
|
| 83 |
+
|
| 84 |
model=model,
|
| 85 |
train_dataset=train_data,
|
| 86 |
eval_dataset=test_data,
|