Upload LlavaForVisionText2Text
Browse files- config.json +1 -0
- generation_config.json +1 -0
- model-00001-of-00004.safetensors +1 -1
- model-00004-of-00004.safetensors +2 -2
- model.safetensors.index.json +2 -1
config.json
CHANGED
@@ -5,6 +5,7 @@
|
|
5 |
"ignore_index": -100,
|
6 |
"image_token_index": 32000,
|
7 |
"model_type": "llava",
|
|
|
8 |
"projector_hidden_act": "gelu",
|
9 |
"text_config": {
|
10 |
"_name_or_path": "mistralai/Mistral-7B-v0.1",
|
|
|
5 |
"ignore_index": -100,
|
6 |
"image_token_index": 32000,
|
7 |
"model_type": "llava",
|
8 |
+
"pad_token_id": 32001,
|
9 |
"projector_hidden_act": "gelu",
|
10 |
"text_config": {
|
11 |
"_name_or_path": "mistralai/Mistral-7B-v0.1",
|
generation_config.json
CHANGED
@@ -2,5 +2,6 @@
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
|
|
5 |
"transformers_version": "4.36.0.dev0"
|
6 |
}
|
|
|
2 |
"_from_model_config": true,
|
3 |
"bos_token_id": 1,
|
4 |
"eos_token_id": 2,
|
5 |
+
"pad_token_id": 32001,
|
6 |
"transformers_version": "4.36.0.dev0"
|
7 |
}
|
model-00001-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4891055040
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ec9698807268456e5a3fa74d657a8bffa5dc161baed731488775454880424011
|
3 |
size 4891055040
|
model-00004-of-00004.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b288c14ec213a3acc8c3eda0611da931cd5d60a2dc6f16550972eb72b4b4fc2f
|
3 |
+
size 933799800
|
model.safetensors.index.json
CHANGED
@@ -1,8 +1,9 @@
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
-
"total_size":
|
4 |
},
|
5 |
"weight_map": {
|
|
|
6 |
"language_model.model.embed_tokens.weight": "model-00001-of-00004.safetensors",
|
7 |
"language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
8 |
"language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|
|
|
1 |
{
|
2 |
"metadata": {
|
3 |
+
"total_size": 15740502016
|
4 |
},
|
5 |
"weight_map": {
|
6 |
+
"language_model.lm_head.weight": "model-00004-of-00004.safetensors",
|
7 |
"language_model.model.embed_tokens.weight": "model-00001-of-00004.safetensors",
|
8 |
"language_model.model.layers.0.input_layernorm.weight": "model-00001-of-00004.safetensors",
|
9 |
"language_model.model.layers.0.mlp.down_proj.weight": "model-00001-of-00004.safetensors",
|