Update config.json
Browse files- config.json +4 -2
config.json
CHANGED
|
@@ -83,7 +83,8 @@
|
|
| 83 |
"hidden_size": 768,
|
| 84 |
"intermediate_size": 3072,
|
| 85 |
"num_attention_heads": 12,
|
| 86 |
-
"num_hidden_layers": 12
|
|
|
|
| 87 |
},
|
| 88 |
"torch_dtype": "float32",
|
| 89 |
"transformers_version": null,
|
|
@@ -164,6 +165,7 @@
|
|
| 164 |
"intermediate_size": 4096,
|
| 165 |
"num_attention_heads": 16,
|
| 166 |
"num_hidden_layers": 24,
|
| 167 |
-
"patch_size": 14
|
|
|
|
| 168 |
}
|
| 169 |
}
|
|
|
|
| 83 |
"hidden_size": 768,
|
| 84 |
"intermediate_size": 3072,
|
| 85 |
"num_attention_heads": 12,
|
| 86 |
+
"num_hidden_layers": 12,
|
| 87 |
+
"projection_dim": 768,
|
| 88 |
},
|
| 89 |
"torch_dtype": "float32",
|
| 90 |
"transformers_version": null,
|
|
|
|
| 165 |
"intermediate_size": 4096,
|
| 166 |
"num_attention_heads": 16,
|
| 167 |
"num_hidden_layers": 24,
|
| 168 |
+
"patch_size": 14,
|
| 169 |
+
"projection_dim": 768,
|
| 170 |
}
|
| 171 |
}
|