Update README.md
Browse files
README.md
CHANGED
|
@@ -25,7 +25,7 @@ This model is a fine-tuned version of the Qwen3-VL-8B-Instruct model using the U
|
|
| 25 |
|
| 26 |
## Training
|
| 27 |
|
| 28 |
-
The model was trained for
|
| 29 |
|
| 30 |
|
| 31 |
## Dataset
|
|
|
|
| 25 |
|
| 26 |
## Training
|
| 27 |
|
| 28 |
+
The model was trained for 4 epochs with a batch size of 8 (4 per device with 8 gradient accumulation steps). LoRA adapters were used for parameter efficient finetuning, targeting both vision and language layers, as well as attention and MLP modules.
|
| 29 |
|
| 30 |
|
| 31 |
## Dataset
|