Fix incorrect bos_token, eos_token, and pad_token ids in config.json (#11)
Browse files- Fix incorrect bos_token, eos_token, and pad_token ids in config.json (925e20f19d50a6d62ea77d55b8b1fa0f6e693cc9)
Co-authored-by: Joshua <[email protected]>
- config.json +3 -3
config.json
CHANGED
|
@@ -7,11 +7,11 @@
|
|
| 7 |
"AutoConfig": "configuration_florence2.Florence2Config",
|
| 8 |
"AutoModelForCausalLM": "modeling_florence2.Florence2ForConditionalGeneration"
|
| 9 |
},
|
| 10 |
-
"bos_token_id":
|
| 11 |
-
"eos_token_id":
|
| 12 |
"ignore_index": -100,
|
| 13 |
"model_type": "florence2",
|
| 14 |
-
"pad_token_id":
|
| 15 |
"projection_dim": 1024,
|
| 16 |
"text_config": {
|
| 17 |
"vocab_size": 51289,
|
|
|
|
| 7 |
"AutoConfig": "configuration_florence2.Florence2Config",
|
| 8 |
"AutoModelForCausalLM": "modeling_florence2.Florence2ForConditionalGeneration"
|
| 9 |
},
|
| 10 |
+
"bos_token_id": 0,
|
| 11 |
+
"eos_token_id": 2,
|
| 12 |
"ignore_index": -100,
|
| 13 |
"model_type": "florence2",
|
| 14 |
+
"pad_token_id": 1,
|
| 15 |
"projection_dim": 1024,
|
| 16 |
"text_config": {
|
| 17 |
"vocab_size": 51289,
|