VPeterV commited on
Commit
4921c22
·
1 Parent(s): 9d82302
Files changed (3) hide show
  1. config.json +92 -0
  2. mm_projector.bin +3 -0
  3. trainer_state.json +0 -0
config.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "SurfaceData/llava-v1.6-mistral-7b-sglang",
3
+ "architectures": [
4
+ "LlavaLlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "freeze_mm_mlp_adapter": false,
11
+ "freeze_mm_vision_resampler": false,
12
+ "hidden_act": "silu",
13
+ "hidden_size": 4096,
14
+ "ignore_index": -100,
15
+ "image_aspect_ratio": "pad",
16
+ "image_crop_resolution": 224,
17
+ "image_grid_pinpoints": [
18
+ [
19
+ 336,
20
+ 672
21
+ ],
22
+ [
23
+ 672,
24
+ 336
25
+ ],
26
+ [
27
+ 672,
28
+ 672
29
+ ],
30
+ [
31
+ 1008,
32
+ 336
33
+ ],
34
+ [
35
+ 336,
36
+ 1008
37
+ ]
38
+ ],
39
+ "image_split_resolution": 224,
40
+ "image_token_index": 32000,
41
+ "initializer_range": 0.02,
42
+ "intermediate_size": 14336,
43
+ "max_position_embeddings": 32768,
44
+ "mm_hidden_size": 1024,
45
+ "mm_patch_merge_type": "flat",
46
+ "mm_projector_lr": null,
47
+ "mm_projector_type": "mlp2x_gelu",
48
+ "mm_resampler_type": null,
49
+ "mm_use_im_patch_token": false,
50
+ "mm_use_im_start_end": false,
51
+ "mm_vision_select_feature": "patch",
52
+ "mm_vision_select_layer": -2,
53
+ "mm_vision_tower": "openai/clip-vit-large-patch14-336",
54
+ "mm_vision_tower_lr": 2e-06,
55
+ "model_type": "llava_llama",
56
+ "num_attention_heads": 32,
57
+ "num_hidden_layers": 32,
58
+ "num_key_value_heads": 8,
59
+ "pretraining_tp": 1,
60
+ "projector_hidden_act": "gelu",
61
+ "rms_norm_eps": 1e-05,
62
+ "rope_scaling": null,
63
+ "rope_theta": 1000000.0,
64
+ "sliding_window": null,
65
+ "text_config": {
66
+ "model_type": "llama"
67
+ },
68
+ "tie_word_embeddings": false,
69
+ "tokenizer_model_max_length": 2048,
70
+ "tokenizer_padding_side": "right",
71
+ "torch_dtype": "bfloat16",
72
+ "transformers_version": "4.37.2",
73
+ "tune_mm_mlp_adapter": true,
74
+ "tune_mm_vision_resampler": false,
75
+ "unfreeze_mm_vision_tower": true,
76
+ "use_cache": true,
77
+ "use_mm_proj": true,
78
+ "vision_config": {
79
+ "hidden_size": 1024,
80
+ "image_size": 336,
81
+ "intermediate_size": 4096,
82
+ "model_type": "clip_vision_model",
83
+ "num_attention_heads": 16,
84
+ "num_hidden_layers": 24,
85
+ "patch_size": 14,
86
+ "projection_dim": 768,
87
+ "vocab_size": 32000
88
+ },
89
+ "vision_feature_layer": -2,
90
+ "vision_feature_select_strategy": "default",
91
+ "vocab_size": 32000
92
+ }
mm_projector.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92f491515928fb2f9b9b921f24694e15a4d3f292f5db1e7479c98d00999d40c8
3
+ size 41961528
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff