zooblastlbz commited on
Commit
a9e1e1a
·
verified ·
1 Parent(s): 17f4e3b

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .editorconfig +18 -0
  2. .gitattributes +137 -35
  3. .gitignore +73 -0
  4. 1 +1 -0
  5. LICENSE +201 -0
  6. README.md +237 -0
  7. all_layers_simulated_rope_qk_dot_product_corrected.pdf +3 -0
  8. all_layers_simulated_rope_qk_dot_product_corrected.png +3 -0
  9. all_layers_simulated_rope_qk_dot_product_corrected.svg +0 -0
  10. attention_heatmap.jpg +0 -0
  11. attention_map/mmpe/attention_heatmap_0.png +0 -0
  12. attention_map/mmpe/attention_heatmap_1.png +0 -0
  13. attention_map/mmpe/attention_heatmap_10.png +0 -0
  14. attention_map/mmpe/attention_heatmap_11.png +0 -0
  15. attention_map/mmpe/attention_heatmap_12.png +0 -0
  16. attention_map/mmpe/attention_heatmap_13.png +0 -0
  17. attention_map/mmpe/attention_heatmap_14.png +0 -0
  18. attention_map/mmpe/attention_heatmap_15.png +0 -0
  19. attention_map/mmpe/attention_heatmap_16.png +0 -0
  20. attention_map/mmpe/attention_heatmap_17.png +0 -0
  21. attention_map/mmpe/attention_heatmap_18.png +0 -0
  22. attention_map/mmpe/attention_heatmap_19.png +0 -0
  23. attention_map/mmpe/attention_heatmap_2.png +0 -0
  24. attention_map/mmpe/attention_heatmap_20.png +0 -0
  25. attention_map/mmpe/attention_heatmap_21.png +0 -0
  26. attention_map/mmpe/attention_heatmap_22.png +0 -0
  27. attention_map/mmpe/attention_heatmap_23.png +0 -0
  28. attention_map/mmpe/attention_heatmap_24.png +0 -0
  29. attention_map/mmpe/attention_heatmap_25.png +0 -0
  30. attention_map/mmpe/attention_heatmap_26.png +0 -0
  31. attention_map/mmpe/attention_heatmap_27.png +0 -0
  32. attention_map/mmpe/attention_heatmap_28.png +0 -0
  33. attention_map/mmpe/attention_heatmap_29.png +0 -0
  34. attention_map/mmpe/attention_heatmap_3.png +0 -0
  35. attention_map/mmpe/attention_heatmap_30.png +0 -0
  36. attention_map/mmpe/attention_heatmap_31.png +0 -0
  37. attention_map/mmpe/attention_heatmap_4.png +0 -0
  38. attention_map/mmpe/attention_heatmap_5.png +0 -0
  39. attention_map/mmpe/attention_heatmap_6.png +0 -0
  40. attention_map/mmpe/attention_heatmap_7.png +0 -0
  41. attention_map/mmpe/attention_heatmap_8.png +0 -0
  42. attention_map/mmpe/attention_heatmap_9.png +0 -0
  43. attention_map/mmpe/attn_layer0_330.png +0 -0
  44. attention_map/mmpe/attn_layer0_336.png +0 -0
  45. attention_map/mmpe/attn_layer0_948.png +0 -0
  46. attention_map/mmpe/attn_layer0_950.png +0 -0
  47. attention_map/mmpe/attn_layer0_952.png +0 -0
  48. attention_map/mmpe/attn_layer0_960.png +0 -0
  49. attention_map/mmpe/attn_layer0_between.png +0 -0
  50. attention_map/mmpe/attn_layer0_each_pair.png +0 -0
.editorconfig ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ root = true
2
+
3
+ # Unix-style newlines with a newline ending every file
4
+ [*]
5
+ end_of_line = lf
6
+ insert_final_newline = true
7
+ trim_trailing_whitespace = true
8
+ charset = utf-8
9
+
10
+ # 4 space indentation
11
+ [*.{py,json}]
12
+ indent_style = space
13
+ indent_size = 4
14
+
15
+ # 2 space indentation
16
+ [*.{md,sh,yaml,yml}]
17
+ indent_style = space
18
+ indent_size = 2
.gitattributes CHANGED
@@ -1,35 +1,137 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # https://git-scm.com/docs/gitattributes
2
+
3
+ # Set the default behavior, in case people don't have core.autocrlf set.
4
+ # https://git-scm.com/docs/gitattributes#_end_of_line_conversion
5
+ * text=auto
6
+
7
+ # common python attributes, taken from https://github.com/alexkaratarakis/gitattributes/blob/710900479a2bedeec7003d381719521ffbb18bf8/Python.gitattributes
8
+ # Source files
9
+ # ============
10
+ *.pxd text diff=python
11
+ *.py text diff=python
12
+ *.py3 text diff=python
13
+ *.pyw text diff=python
14
+ *.pyx text diff=python
15
+ *.pyz text diff=python
16
+ *.pyi text diff=python
17
+
18
+ # Binary files
19
+ # ============
20
+ *.db binary
21
+ *.p binary
22
+ *.pkl binary
23
+ *.pickle binary
24
+ *.pyc binary export-ignore
25
+ *.pyo binary export-ignore
26
+ *.pyd binary
27
+
28
+ # Jupyter notebook
29
+ *.ipynb text eol=lf
30
+ all_layers_simulated_rope_qk_dot_product_corrected.pdf filter=lfs diff=lfs merge=lfs -text
31
+ all_layers_simulated_rope_qk_dot_product_corrected.png filter=lfs diff=lfs merge=lfs -text
32
+ attention_map/mmpe/attn_layer0_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
33
+ attention_map/mmpe/attn_layer10_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
34
+ attention_map/mmpe/attn_layer11_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
35
+ attention_map/mmpe/attn_layer12_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
36
+ attention_map/mmpe/attn_layer13_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
37
+ attention_map/mmpe/attn_layer14_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
38
+ attention_map/mmpe/attn_layer15_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
39
+ attention_map/mmpe/attn_layer16_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
40
+ attention_map/mmpe/attn_layer17_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
41
+ attention_map/mmpe/attn_layer18_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
42
+ attention_map/mmpe/attn_layer19_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
43
+ attention_map/mmpe/attn_layer1_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
44
+ attention_map/mmpe/attn_layer20_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
45
+ attention_map/mmpe/attn_layer21_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
46
+ attention_map/mmpe/attn_layer22_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
47
+ attention_map/mmpe/attn_layer23_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
48
+ attention_map/mmpe/attn_layer24_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
49
+ attention_map/mmpe/attn_layer25_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
50
+ attention_map/mmpe/attn_layer26_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
51
+ attention_map/mmpe/attn_layer27_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
52
+ attention_map/mmpe/attn_layer28_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
53
+ attention_map/mmpe/attn_layer29_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
54
+ attention_map/mmpe/attn_layer2_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
55
+ attention_map/mmpe/attn_layer30_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
56
+ attention_map/mmpe/attn_layer31_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
57
+ attention_map/mmpe/attn_layer3_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
58
+ attention_map/mmpe/attn_layer4_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
59
+ attention_map/mmpe/attn_layer5_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
60
+ attention_map/mmpe/attn_layer6_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
61
+ attention_map/mmpe/attn_layer7_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
62
+ attention_map/mmpe/attn_layer8_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
63
+ attention_map/mmpe/attn_layer9_pos330_comparison.png filter=lfs diff=lfs merge=lfs -text
64
+ attention_map/mmpe/position_950_highlight.png filter=lfs diff=lfs merge=lfs -text
65
+ attention_map/mmpe/position_ids_overlay.png filter=lfs diff=lfs merge=lfs -text
66
+ attention_map/mmpe_two/attn_layer0_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
67
+ attention_map/mmpe_two/attn_layer10_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
68
+ attention_map/mmpe_two/attn_layer11_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
69
+ attention_map/mmpe_two/attn_layer12_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
70
+ attention_map/mmpe_two/attn_layer13_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
71
+ attention_map/mmpe_two/attn_layer14_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
72
+ attention_map/mmpe_two/attn_layer15_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
73
+ attention_map/mmpe_two/attn_layer16_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
74
+ attention_map/mmpe_two/attn_layer17_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
75
+ attention_map/mmpe_two/attn_layer18_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
76
+ attention_map/mmpe_two/attn_layer19_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
77
+ attention_map/mmpe_two/attn_layer1_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
78
+ attention_map/mmpe_two/attn_layer20_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
79
+ attention_map/mmpe_two/attn_layer21_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
80
+ attention_map/mmpe_two/attn_layer22_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
81
+ attention_map/mmpe_two/attn_layer23_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
82
+ attention_map/mmpe_two/attn_layer24_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
83
+ attention_map/mmpe_two/attn_layer25_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
84
+ attention_map/mmpe_two/attn_layer26_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
85
+ attention_map/mmpe_two/attn_layer27_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
86
+ attention_map/mmpe_two/attn_layer28_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
87
+ attention_map/mmpe_two/attn_layer29_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
88
+ attention_map/mmpe_two/attn_layer2_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
89
+ attention_map/mmpe_two/attn_layer30_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
90
+ attention_map/mmpe_two/attn_layer31_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
91
+ attention_map/mmpe_two/attn_layer3_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
92
+ attention_map/mmpe_two/attn_layer4_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
93
+ attention_map/mmpe_two/attn_layer5_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
94
+ attention_map/mmpe_two/attn_layer6_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
95
+ attention_map/mmpe_two/attn_layer7_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
96
+ attention_map/mmpe_two/attn_layer8_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
97
+ attention_map/mmpe_two/attn_layer9_each_pair_comparison.png filter=lfs diff=lfs merge=lfs -text
98
+ attention_map/without_mmpe/position_950_highlight.png filter=lfs diff=lfs merge=lfs -text
99
+ attention_map/without_mmpe/position_ids_overlay.png filter=lfs diff=lfs merge=lfs -text
100
+ docs/jobs.mp4 filter=lfs diff=lfs merge=lfs -text
101
+ docs/ov_chat_images/chat_results.png filter=lfs diff=lfs merge=lfs -text
102
+ docs/ov_chat_images/example1_tree.png filter=lfs diff=lfs merge=lfs -text
103
+ docs/ov_chat_images/example2_dog.jpg filter=lfs diff=lfs merge=lfs -text
104
+ llava/model/multimodal_encoder/dev_eva_clip/eva_clip/bpe_simple_vocab_16e6.txt.gz filter=lfs diff=lfs merge=lfs -text
105
+ per_head/layer_0_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
106
+ per_head/layer_10_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
107
+ per_head/layer_11_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
108
+ per_head/layer_12_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
109
+ per_head/layer_13_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
110
+ per_head/layer_14_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
111
+ per_head/layer_15_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
112
+ per_head/layer_16_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
113
+ per_head/layer_17_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
114
+ per_head/layer_18_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
115
+ per_head/layer_19_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
116
+ per_head/layer_1_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
117
+ per_head/layer_20_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
118
+ per_head/layer_21_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
119
+ per_head/layer_22_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
120
+ per_head/layer_23_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
121
+ per_head/layer_24_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
122
+ per_head/layer_25_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
123
+ per_head/layer_26_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
124
+ per_head/layer_27_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
125
+ per_head/layer_28_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
126
+ per_head/layer_29_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
127
+ per_head/layer_2_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
128
+ per_head/layer_30_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
129
+ per_head/layer_31_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
130
+ per_head/layer_3_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
131
+ per_head/layer_4_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
132
+ per_head/layer_5_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
133
+ per_head/layer_6_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
134
+ per_head/layer_7_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
135
+ per_head/layer_8_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
136
+ per_head/layer_9_per_head_simulated_rope_qk_dot_product.pdf filter=lfs diff=lfs merge=lfs -text
137
+ playground/demo/xU25MMA2N4aVtYay.mp4 filter=lfs diff=lfs merge=lfs -text
.gitignore ADDED
@@ -0,0 +1,73 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Python
2
+ __pycache__
3
+ *.pyc
4
+ *.egg-info
5
+ dist
6
+
7
+ # Log
8
+ *.log
9
+ *.log.*
10
+ # *.json
11
+ # *.jsonl
12
+
13
+ # Data
14
+ !**/alpaca-data-conversation.json
15
+ # Editor
16
+ .idea
17
+ *.swp
18
+ .vscode
19
+
20
+ # Other
21
+ .DS_Store
22
+ wandb
23
+ output
24
+ llavavid
25
+
26
+ checkpoints
27
+ project_checkpoints
28
+ debug_checkpoints
29
+ playground/data
30
+ playground/cc3m_llava34b_cap
31
+ ckpts*
32
+
33
+ .ipynb_checkpoints
34
+ chunyl_scripts
35
+ *.ipynb
36
+
37
+ # DevContainer
38
+ !.devcontainer/*
39
+
40
+ # Demo
41
+ serve_images/
42
+ notebooks/
43
+ logs
44
+ scripts/dist_*
45
+ logs/
46
+ submissions/
47
+ cn_scripts/
48
+ internal_project_checkpoints/
49
+ work_dirs
50
+ scripts/i18n/*
51
+ playground/.nfs028b000000010add00000001
52
+ HIP
53
+ playground/.nfs028b0000017bff2c00000012
54
+ scripts/qwen
55
+ scripts/vicuna
56
+ scripts/mistral
57
+ scripts/baseline_rep
58
+ scripts/cn_boli01_hl
59
+ scripts/cn_boli01_lf
60
+ scripts/cn_lf
61
+ scripts/cn_lq
62
+ scripts/cn_yg
63
+ scripts/cn_yg_hao
64
+ scripts/eva_encoder
65
+ scripts/i18n
66
+ scripts/i18n_higher_res
67
+ scripts/multi-images
68
+ scratchpad
69
+ build/
70
+ playground/*.json
71
+ mlx_configs/
72
+ data_processing/
73
+ # demo/
1 ADDED
@@ -0,0 +1 @@
 
 
1
+ nohup: ignoring input
LICENSE ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Apache License
2
+ Version 2.0, January 2004
3
+ http://www.apache.org/licenses/
4
+
5
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
6
+
7
+ 1. Definitions.
8
+
9
+ "License" shall mean the terms and conditions for use, reproduction,
10
+ and distribution as defined by Sections 1 through 9 of this document.
11
+
12
+ "Licensor" shall mean the copyright owner or entity authorized by
13
+ the copyright owner that is granting the License.
14
+
15
+ "Legal Entity" shall mean the union of the acting entity and all
16
+ other entities that control, are controlled by, or are under common
17
+ control with that entity. For the purposes of this definition,
18
+ "control" means (i) the power, direct or indirect, to cause the
19
+ direction or management of such entity, whether by contract or
20
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
21
+ outstanding shares, or (iii) beneficial ownership of such entity.
22
+
23
+ "You" (or "Your") shall mean an individual or Legal Entity
24
+ exercising permissions granted by this License.
25
+
26
+ "Source" form shall mean the preferred form for making modifications,
27
+ including but not limited to software source code, documentation
28
+ source, and configuration files.
29
+
30
+ "Object" form shall mean any form resulting from mechanical
31
+ transformation or translation of a Source form, including but
32
+ not limited to compiled object code, generated documentation,
33
+ and conversions to other media types.
34
+
35
+ "Work" shall mean the work of authorship, whether in Source or
36
+ Object form, made available under the License, as indicated by a
37
+ copyright notice that is included in or attached to the work
38
+ (an example is provided in the Appendix below).
39
+
40
+ "Derivative Works" shall mean any work, whether in Source or Object
41
+ form, that is based on (or derived from) the Work and for which the
42
+ editorial revisions, annotations, elaborations, or other modifications
43
+ represent, as a whole, an original work of authorship. For the purposes
44
+ of this License, Derivative Works shall not include works that remain
45
+ separable from, or merely link (or bind by name) to the interfaces of,
46
+ the Work and Derivative Works thereof.
47
+
48
+ "Contribution" shall mean any work of authorship, including
49
+ the original version of the Work and any modifications or additions
50
+ to that Work or Derivative Works thereof, that is intentionally
51
+ submitted to Licensor for inclusion in the Work by the copyright owner
52
+ or by an individual or Legal Entity authorized to submit on behalf of
53
+ the copyright owner. For the purposes of this definition, "submitted"
54
+ means any form of electronic, verbal, or written communication sent
55
+ to the Licensor or its representatives, including but not limited to
56
+ communication on electronic mailing lists, source code control systems,
57
+ and issue tracking systems that are managed by, or on behalf of, the
58
+ Licensor for the purpose of discussing and improving the Work, but
59
+ excluding communication that is conspicuously marked or otherwise
60
+ designated in writing by the copyright owner as "Not a Contribution."
61
+
62
+ "Contributor" shall mean Licensor and any individual or Legal Entity
63
+ on behalf of whom a Contribution has been received by Licensor and
64
+ subsequently incorporated within the Work.
65
+
66
+ 2. Grant of Copyright License. Subject to the terms and conditions of
67
+ this License, each Contributor hereby grants to You a perpetual,
68
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
69
+ copyright license to reproduce, prepare Derivative Works of,
70
+ publicly display, publicly perform, sublicense, and distribute the
71
+ Work and such Derivative Works in Source or Object form.
72
+
73
+ 3. Grant of Patent License. Subject to the terms and conditions of
74
+ this License, each Contributor hereby grants to You a perpetual,
75
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
76
+ (except as stated in this section) patent license to make, have made,
77
+ use, offer to sell, sell, import, and otherwise transfer the Work,
78
+ where such license applies only to those patent claims licensable
79
+ by such Contributor that are necessarily infringed by their
80
+ Contribution(s) alone or by combination of their Contribution(s)
81
+ with the Work to which such Contribution(s) was submitted. If You
82
+ institute patent litigation against any entity (including a
83
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
84
+ or a Contribution incorporated within the Work constitutes direct
85
+ or contributory patent infringement, then any patent licenses
86
+ granted to You under this License for that Work shall terminate
87
+ as of the date such litigation is filed.
88
+
89
+ 4. Redistribution. You may reproduce and distribute copies of the
90
+ Work or Derivative Works thereof in any medium, with or without
91
+ modifications, and in Source or Object form, provided that You
92
+ meet the following conditions:
93
+
94
+ (a) You must give any other recipients of the Work or
95
+ Derivative Works a copy of this License; and
96
+
97
+ (b) You must cause any modified files to carry prominent notices
98
+ stating that You changed the files; and
99
+
100
+ (c) You must retain, in the Source form of any Derivative Works
101
+ that You distribute, all copyright, patent, trademark, and
102
+ attribution notices from the Source form of the Work,
103
+ excluding those notices that do not pertain to any part of
104
+ the Derivative Works; and
105
+
106
+ (d) If the Work includes a "NOTICE" text file as part of its
107
+ distribution, then any Derivative Works that You distribute must
108
+ include a readable copy of the attribution notices contained
109
+ within such NOTICE file, excluding those notices that do not
110
+ pertain to any part of the Derivative Works, in at least one
111
+ of the following places: within a NOTICE text file distributed
112
+ as part of the Derivative Works; within the Source form or
113
+ documentation, if provided along with the Derivative Works; or,
114
+ within a display generated by the Derivative Works, if and
115
+ wherever such third-party notices normally appear. The contents
116
+ of the NOTICE file are for informational purposes only and
117
+ do not modify the License. You may add Your own attribution
118
+ notices within Derivative Works that You distribute, alongside
119
+ or as an addendum to the NOTICE text from the Work, provided
120
+ that such additional attribution notices cannot be construed
121
+ as modifying the License.
122
+
123
+ You may add Your own copyright statement to Your modifications and
124
+ may provide additional or different license terms and conditions
125
+ for use, reproduction, or distribution of Your modifications, or
126
+ for any such Derivative Works as a whole, provided Your use,
127
+ reproduction, and distribution of the Work otherwise complies with
128
+ the conditions stated in this License.
129
+
130
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
131
+ any Contribution intentionally submitted for inclusion in the Work
132
+ by You to the Licensor shall be under the terms and conditions of
133
+ this License, without any additional terms or conditions.
134
+ Notwithstanding the above, nothing herein shall supersede or modify
135
+ the terms of any separate license agreement you may have executed
136
+ with Licensor regarding such Contributions.
137
+
138
+ 6. Trademarks. This License does not grant permission to use the trade
139
+ names, trademarks, service marks, or product names of the Licensor,
140
+ except as required for reasonable and customary use in describing the
141
+ origin of the Work and reproducing the content of the NOTICE file.
142
+
143
+ 7. Disclaimer of Warranty. Unless required by applicable law or
144
+ agreed to in writing, Licensor provides the Work (and each
145
+ Contributor provides its Contributions) on an "AS IS" BASIS,
146
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
147
+ implied, including, without limitation, any warranties or conditions
148
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
149
+ PARTICULAR PURPOSE. You are solely responsible for determining the
150
+ appropriateness of using or redistributing the Work and assume any
151
+ risks associated with Your exercise of permissions under this License.
152
+
153
+ 8. Limitation of Liability. In no event and under no legal theory,
154
+ whether in tort (including negligence), contract, or otherwise,
155
+ unless required by applicable law (such as deliberate and grossly
156
+ negligent acts) or agreed to in writing, shall any Contributor be
157
+ liable to You for damages, including any direct, indirect, special,
158
+ incidental, or consequential damages of any character arising as a
159
+ result of this License or out of the use or inability to use the
160
+ Work (including but not limited to damages for loss of goodwill,
161
+ work stoppage, computer failure or malfunction, or any and all
162
+ other commercial damages or losses), even if such Contributor
163
+ has been advised of the possibility of such damages.
164
+
165
+ 9. Accepting Warranty or Additional Liability. While redistributing
166
+ the Work or Derivative Works thereof, You may choose to offer,
167
+ and charge a fee for, acceptance of support, warranty, indemnity,
168
+ or other liability obligations and/or rights consistent with this
169
+ License. However, in accepting such obligations, You may act only
170
+ on Your own behalf and on Your sole responsibility, not on behalf
171
+ of any other Contributor, and only if You agree to indemnify,
172
+ defend, and hold each Contributor harmless for any liability
173
+ incurred by, or claims asserted against, such Contributor by reason
174
+ of your accepting any such warranty or additional liability.
175
+
176
+ END OF TERMS AND CONDITIONS
177
+
178
+ APPENDIX: How to apply the Apache License to your work.
179
+
180
+ To apply the Apache License to your work, attach the following
181
+ boilerplate notice, with the fields enclosed by brackets "[]"
182
+ replaced with your own identifying information. (Don't include
183
+ the brackets!) The text should be enclosed in the appropriate
184
+ comment syntax for the file format. We also recommend that a
185
+ file or class name and description of purpose be included on the
186
+ same "printed page" as the copyright notice for easier
187
+ identification within third-party archives.
188
+
189
+ Copyright [yyyy] [name of copyright owner]
190
+
191
+ Licensed under the Apache License, Version 2.0 (the "License");
192
+ you may not use this file except in compliance with the License.
193
+ You may obtain a copy of the License at
194
+
195
+ http://www.apache.org/licenses/LICENSE-2.0
196
+
197
+ Unless required by applicable law or agreed to in writing, software
198
+ distributed under the License is distributed on an "AS IS" BASIS,
199
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
200
+ See the License for the specific language governing permissions and
201
+ limitations under the License.
README.md ADDED
@@ -0,0 +1,237 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ <p align="center" width="100%">
2
+ <img src="https://i.postimg.cc/pL17YtG4/WX20240508-220230-2x.png" width="80%" height="80%">
3
+ </p>
4
+
5
+ # LLaVA-NeXT: Open Large Multimodal Models
6
+ [![Static Badge](https://img.shields.io/badge/llava_video-paper-green)](http://arxiv.org/abs/2410.02713)
7
+ [![Static Badge](https://img.shields.io/badge/llava_onevision-paper-green)](https://arxiv.org/abs/2408.03326)
8
+ [![llava_next-blog](https://img.shields.io/badge/llava_next-blog-green)](https://llava-vl.github.io/blog/)
9
+
10
+ [![llava_onevision-demo](https://img.shields.io/badge/llava_onevision-demo-red)](https://llava-onevision.lmms-lab.com/)
11
+ [![llava_next-video_demo](https://img.shields.io/badge/llava_video-demo-red)](https://huggingface.co/spaces/WildVision/vision-arena)
12
+ [![llava_next-interleave_demo](https://img.shields.io/badge/llava_next-interleave_demo-red)](https://huggingface.co/spaces/lmms-lab/LLaVA-NeXT-Interleave-Demo)
13
+ [![Openbayes Demo](https://img.shields.io/static/v1?label=Demo&message=OpenBayes%E8%B4%9D%E5%BC%8F%E8%AE%A1%E7%AE%97&color=green)](https://openbayes.com/console/public/tutorials/gW0ng9jKXfO)
14
+
15
+ [![llava_video-checkpoints](https://img.shields.io/badge/llava_video-checkpoints-blue)](https://huggingface.co/collections/lmms-lab/llava-next-video-661e86f5e8dabc3ff793c944)
16
+ [![llava_onevision-checkpoints](https://img.shields.io/badge/llava_onevision-checkpoints-blue)](https://huggingface.co/collections/lmms-lab/llava-onevision-66a259c3526e15166d6bba37)
17
+ [![llava_next-interleave_checkpoints](https://img.shields.io/badge/llava_next-interleave_checkpoints-blue)](https://huggingface.co/collections/lmms-lab/llava-next-interleave-66763c55c411b340b35873d1)
18
+ [![llava_next-image_checkpoints](https://img.shields.io/badge/llava_next-image_checkpoints-blue)](https://huggingface.co/lmms-lab)
19
+
20
+ ## Release Notes
21
+
22
+ - **[2024/10/04] 🔥 LLaVA-Video** (formerly LLaVA-NeXT-Video) has undergone a major upgrade! We are excited to release **LLaVA-Video-178K**, a high-quality synthetic dataset for video instruction tuning. This dataset includes:
23
+
24
+ - 178,510 caption entries
25
+ - 960,792 open-ended Q&A pairs
26
+ - 196,198 multiple-choice Q&A items
27
+
28
+ Along with this, we’re also releasing the **LLaVA-Video 7B/72B models**, which deliver competitive performance on the latest video benchmarks, including [Video-MME](https://video-mme.github.io/home_page.html#leaderboard), [LongVideoBench](https://longvideobench.github.io/), and [Dream-1K](https://tarsier-vlm.github.io/).
29
+
30
+ 📄 **Explore more**:
31
+ - [LLaVA-Video-178K Dataset](https://huggingface.co/datasets/lmms-lab/LLaVA-Video-178K): Download the dataset.
32
+ - [LLaVA-Video Models](https://huggingface.co/collections/lmms-lab/llava-video-661e86f5e8dabc3ff793c944): Access model checkpoints.
33
+ - [Paper](http://arxiv.org/abs/2410.02713): Detailed information about LLaVA-Video.
34
+ - [LLaVA-Video Documentation](https://github.com/LLaVA-VL/LLaVA-NeXT/blob/main/docs/LLaVA_Video_1003.md): Guidance on training, inference and evaluation.
35
+
36
+ - [2024/09/13] 🔥 **🚀 [LLaVA-OneVision-Chat](docs/LLaVA_OneVision_Chat.md)**. The new LLaVA-OV-Chat (7B/72B) significantly improves the chat experience of LLaVA-OV. 📄
37
+
38
+ ![](docs/ov_chat_images/chat_results.png)
39
+
40
+ - [2024/08/06] 🔥 **🚀 [LLaVA-OneVision (OV)](https://llava-vl.github.io/blog/2024-08-05-llava-onevision/)!** The new LLaVA-OV models (0.5B/7B/72B) achieve new state-of-the-art performance across single-image, multi-image, and video benchmarks, sometimes rivaling top commercial models on 47 diverse benchmarks. 📄 Explore More:
41
+ * [[Paper]](https://arxiv.org/abs/2408.03326): In-depth insights, new emegerging scenarios, ie, strong video understadning through task transfer from images.
42
+ * [[LLaVA-OV Doc]](https://github.com/LLaVA-VL/LLaVA-NeXT/blob/main/docs/LLaVA_OneVision.md): Model inference and evaluation guidance.
43
+ * [[Scripts]](https://github.com/LLaVA-VL/LLaVA-NeXT/blob/main/scripts/train): Start training models on your single-image/multi-image/video data.
44
+
45
+ - [2024/07/16] 🔥 **LLaVA-NeXT-Video** has been upgraded. The new 32B model achieves the best open-source performance on several video benchmarks, including [Video-MME](https://video-mme.github.io/home_page.html#leaderboard). Please refer to [this page](docs/LLaVA-NeXT-Video_0716.md) for details, refer to [llava_next-video_demo](https://huggingface.co/spaces/WildVision/vision-arena) for demo.
46
+
47
+
48
+ - [2024/06/23] 🔥 **LLaVA-NeXT-Interleave** is released. We utilize image-text interleaved format to unify multi-image, video, and 3D tasks in one LLM and achieve **SoTA** performance on a wide range of benchmarks. Check out [paper](https://arxiv.org/pdf/2407.07895), [blog](https://llava-vl.github.io/blog/2024-06-16-llava-next-interleave/), and [checkpoints](https://huggingface.co/collections/lmms-lab/llava-next-interleave-66763c55c411b340b35873d1) to see new capabilities and improved performance! We have released 0.5b, 7b, and 7b-dpo models.
49
+ * An all-round LLM for multi-image, video, and 3D with strong performance \[[demo](https://huggingface.co/spaces/lmms-lab/LLaVA-NeXT-Interleave-Demo)\]
50
+ * Construct interleave training data [**M4-Instruct**](https://huggingface.co/datasets/lmms-lab/M4-Instruct-Data)
51
+ * Construct multi-image benchmark [**LLaVA-Interleave Bench**](https://huggingface.co/datasets/lmms-lab/LLaVA-NeXT-Interleave-Bench)
52
+
53
+
54
+ - [2024/05/25] 🔥 Wondering "[What Else Influences Visual Instruction Tuning Beyond Data?](https://llava-vl.github.io/blog/2024-05-25-llava-next-ablations/)" Our new [blog](https://llava-vl.github.io/blog/2024-05-25-llava-next-ablations/) summarizes empirical explorations to ablate the various design choices in improving LMMs except instruct data itself. Meanwhile, open-source the recapioned high-quality data using LLaVA-NeXT-34B on [[COCO]](https://huggingface.co/datasets/lmms-lab/LLaVA-ReCap-118K) [[LCS]](https://huggingface.co/datasets/lmms-lab/LLaVA-ReCap-558K) [[CC3M]](https://huggingface.co/datasets/lmms-lab/LLaVA-ReCap-CC3M).
55
+ * Architectures (LMM & Vision Encoder)
56
+ * Visual Representations (Resolution & # Tokens)
57
+ * Training Strategies (High-quality data & Trainable modules)
58
+
59
+ - [2024/05/10] 🔥 **LLaVA-NeXT** (Stronger) models are released, with support of stronger LMM inlcuding LLama-3 (8B) and Qwen-1.5 (72B/110B) Check out [[blog](https://llava-vl.github.io/blog/2024-05-10-llava-next-stronger-llms/)] and [[checkpoints](https://huggingface.co/lmms-lab)] to see improved performance!
60
+ - [2024/05/10] 🔥 **LLaVA-NeXT** (Video) is released. The image-only-trained LLaVA-NeXT model is surprisingly strong on video tasks with zero-shot modality transfer. DPO training with AI feedback on videos can yield significant improvement. [[Blog](https://llava-vl.github.io/blog/2024-04-30-llava-next-video/)], [[checkpoints](https://huggingface.co/collections/lmms-lab/llava-next-video-661e86f5e8dabc3ff793c944)] and [[sglang](https://github.com/sgl-project/sglang)]
61
+ - [2024/01/30] 🔥 **LLaVA-NeXT** is out! With additional scaling to LLaVA-1.5, LLaVA-NeXT-34B outperforms Gemini Pro on some benchmarks. It can now process 4x more pixels and perform more tasks/applications than before. Check out the [blog post](https://llava-vl.github.io/blog/2024-01-30-llava-next/), and explore the [demo](https://llava.hliu.cc/)! Models are available in [Model Zoo](https://github.com/haotian-liu/LLaVA/blob/main/docs/MODEL_ZOO.md). Training/eval data and scripts coming soon.
62
+ <details>
63
+ <summary>More</summary>
64
+
65
+ - [2024/03/10] 🔥 Releasing **LMMs-Eval**, a highly efficient evaluation pipeline we used when developing LLaVA-NeXT. It supports the evaluation of LMMs on dozens of public datasets and allows new dataset onboarding, making the dev of new LMMs much faster. [[Blog](https://lmms-lab.github.io/lmms-eval-blog/lmms-eval-0.1/)] [[Codebase](https://github.com/EvolvingLMMs-Lab/lmms-eval)]
66
+
67
+ - [2023/11/10] [LLaVA-Plus](https://llava-vl.github.io/llava-plus/) is released: Learning to Use Tools for Creating Multimodal Agents, with LLaVA-Plus (LLaVA that Plug and Learn to Use Skills). [[Project Page](https://llava-vl.github.io/llava-plus/)] [[Demo](https://llavaplus.ngrok.io/)] [[Code](https://github.com/LLaVA-VL/LLaVA-Plus-Codebase)] [[Paper](https://arxiv.org/abs/2311.05437)]
68
+ - [2023/11/02] [LLaVA-Interactive](https://llava-vl.github.io/llava-interactive/) is released: Experience the future of human-AI multimodal interaction with an all-in-one demo for Image Chat, Segmentation, Generation and Editing. [[Project Page](https://llava-vl.github.io/llava-interactive/)] [[Demo](https://llavainteractive.ngrok.io/)] [[Code](https://github.com/LLaVA-VL/LLaVA-Interactive-Demo)] [[Paper](https://arxiv.org/abs/2311.00571)]
69
+ - [2023/10/26] 🔥 LLaVA-1.5 with LoRA achieves comparable performance as full-model finetuning, with a reduced GPU RAM requirement ([ckpts](https://github.com/haotian-liu/LLaVA/blob/main/docs/MODEL_ZOO.md#llava-v15), [script](https://github.com/haotian-liu/LLaVA#train)). We also provide a [doc](https://github.com/haotian-liu/LLaVA/blob/main/docs/Finetune_Custom_Data.md) on how to finetune LLaVA-1.5 on your own dataset with LoRA.
70
+ - [2023/10/12] Check out the Korean LLaVA (Ko-LLaVA), created by ETRI, who has generously supported our research! [[🤗 Demo](https://huggingface.co/spaces/etri-vilab/Ko-LLaVA)]
71
+ - [2023/10/05] 🔥 LLaVA-1.5 is out! Achieving SoTA on 11 benchmarks, with just simple modifications to the original LLaVA, utilizes all public data, completes training in ~1 day on a single 8-A100 node, and surpasses methods like Qwen-VL-Chat that use billion-scale data. Check out the [technical report](https://arxiv.org/abs/2310.03744), and explore the [demo](https://llava.hliu.cc/)! Models are available in [Model Zoo](https://github.com/haotian-liu/LLaVA/blob/main/docs/MODEL_ZOO.md). The training data and scripts of LLaVA-1.5 are released [here](https://github.com/haotian-liu/LLaVA#train), and evaluation scripts are released [here](https://github.com/haotian-liu/LLaVA/blob/main/docs/Evaluation.md)!
72
+ - [2023/09/26] LLaVA is improved with reinforcement learning from human feedback (RLHF) to improve fact grounding and reduce hallucination. Check out the new SFT and RLHF checkpoints at project [[LLavA-RLHF]](https://llava-rlhf.github.io/)
73
+ - [2023/09/22] [LLaVA](https://arxiv.org/abs/2304.08485) is accepted by NeurIPS 2023 as **oral presentation**, and [LLaVA-Med](https://arxiv.org/abs/2306.00890) is accepted by NeurIPS 2023 Datasets and Benchmarks Track as **spotlight presentation**.
74
+ - [2023/11/06] Support **Intel** dGPU and CPU platforms. [More details here.](https://github.com/haotian-liu/LLaVA/tree/intel/docs/intel)
75
+ - [2023/10/12] LLaVA is now supported in [llama.cpp](https://github.com/ggerganov/llama.cpp/pull/3436) with 4-bit / 5-bit quantization support!
76
+ - [2023/10/11] The training data and scripts of LLaVA-1.5 are released [here](https://github.com/haotian-liu/LLaVA#train), and evaluation scripts are released [here](https://github.com/haotian-liu/LLaVA/blob/main/docs/Evaluation.md)!
77
+ - [2023/10/10] [Roboflow Deep Dive](https://blog.roboflow.com/first-impressions-with-llava-1-5/): First Impressions with LLaVA-1.5.
78
+ - [2023/09/20] We summarize our empirical study of training 33B and 65B LLaVA models in a [note](https://arxiv.org/abs/2309.09958). Further, if you are interested in the comprehensive review, evolution and trend of multimodal foundation models, please check out our recent survey paper [``Multimodal Foundation Models: From Specialists to General-Purpose Assistants''.](https://arxiv.org/abs/2309.10020)
79
+ <p align="center">
80
+ <img src="https://github.com/Computer-Vision-in-the-Wild/CVinW_Readings/blob/main/images/mfm_evolution.jpeg?raw=true" width=50%/>
81
+ </p>
82
+
83
+ - [2023/07/19] 🔥 We release a major upgrade, including support for LLaMA-2, LoRA training, 4-/8-bit inference, higher resolution (336x336), and a lot more. We release [LLaVA Bench](https://github.com/haotian-liu/LLaVA/blob/main/docs/LLaVA_Bench.md) for benchmarking open-ended visual chat with results from Bard and Bing-Chat. We also support and verify training with RTX 3090 and RTX A6000. Check out [LLaVA-from-LLaMA-2](https://github.com/haotian-liu/LLaVA/blob/main/docs/LLaVA_from_LLaMA2.md), and our [model zoo](https://github.com/haotian-liu/LLaVA/blob/main/docs/MODEL_ZOO.md)!
84
+ - [2023/06/26] [CVPR 2023 Tutorial](https://vlp-tutorial.github.io/) on **Large Multimodal Models: Towards Building and Surpassing Multimodal GPT-4**! Please check out [[Slides](https://datarelease.blob.core.windows.net/tutorial/vision_foundation_models_2023/slides/Chunyuan_cvpr2023_tutorial_lmm.pdf)] [[Notes](https://arxiv.org/abs/2306.14895)] [[YouTube](https://youtu.be/mkI7EPD1vp8)] [[Bilibli](https://www.bilibili.com/video/BV1Ng4y1T7v3/)].
85
+ - [2023/06/11] We released the preview for the most requested feature: DeepSpeed and LoRA support! Please see documentations [here](./docs/LoRA.md).
86
+ - [2023/06/01] We released **LLaVA-Med: Large Language and Vision Assistant for Biomedicine**, a step towards building biomedical domain large language and vision models with GPT-4 level capabilities. Checkout the [paper](https://arxiv.org/abs/2306.00890) and [page](https://github.com/microsoft/LLaVA-Med).
87
+ - [2023/05/06] We are releasing [LLaVA-Lighting-MPT-7B-preview](https://huggingface.co/liuhaotian/LLaVA-Lightning-MPT-7B-preview), based on MPT-7B-Chat! See [here](#LLaVA-MPT-7b) for more details.
88
+ - [2023/05/02] 🔥 We are releasing LLaVA-Lighting! Train a lite, multimodal GPT-4 with just $40 in 3 hours! See [here](#train-llava-lightning) for more details.
89
+ - [2023/04/27] Thanks to the community effort, LLaVA-13B with 4-bit quantization allows you to run on a GPU with as few as 12GB VRAM! Try it out [here](https://github.com/oobabooga/text-generation-webui/tree/main/extensions/llava).
90
+ - [2023/04/17] 🔥 We released **LLaVA: Large Language and Vision Assistant**. We propose visual instruction tuning, towards building large language and vision models with GPT-4 level capabilities. Checkout the [paper](https://arxiv.org/abs/2304.08485) and [demo](https://llava.hliu.cc/).
91
+
92
+ </details>
93
+
94
+ <!-- <a href="https://llava.hliu.cc/"><img src="assets/demo.gif" width="70%"></a> -->
95
+
96
+ **Usage and License Notices**: This project utilizes certain datasets and checkpoints that are subject to their respective original licenses. Users must comply with all terms and conditions of these original licenses, including but not limited to the [OpenAI Terms of Use](https://openai.com/policies/terms-of-use) for the dataset and the specific licenses for base language models for checkpoints trained using the dataset (e.g. [Llama-1/2 community license](https://ai.meta.com/llama/license/) for LLaMA-2 and Vicuna-v1.5, [Tongyi Qianwen RESEARCH LICENSE AGREEMENT](https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat/blob/main/LICENSE) and [Llama-3 Research License](https://llama.meta.com/llama3/license/)). This project does not impose any additional constraints beyond those stipulated in the original licenses. Furthermore, users are reminded to ensure that their use of the dataset and checkpoints is in compliance with all applicable laws and regulations.
97
+
98
+ ## Models & Scripts
99
+
100
+ ### Installation
101
+
102
+ #### 1. **Clone this repository and navigate to the LLaVA folder:**
103
+ ```bash
104
+ git clone https://github.com/LLaVA-VL/LLaVA-NeXT
105
+ cd LLaVA-NeXT
106
+ ```
107
+
108
+ #### 2. **Install the inference package:**
109
+ ```bash
110
+ conda create -n llava python=3.10 -y
111
+ conda activate llava
112
+ pip install --upgrade pip # Enable PEP 660 support.
113
+ pip install -e ".[train]"
114
+ ```
115
+
116
+ ### Project Navigation
117
+ Please checkout the following page for more inference & evaluation details.
118
+
119
+ #### - **LLaVA-OneVision: Easy Task Transfer**
120
+ - [LLaVA-OneVision]([./docs/LLaVA-NeXT.md](https://github.com/LLaVA-VL/LLaVA-NeXT/blob/main/docs/LLaVA_OneVision.md)): for demo inference. The evaluation code is in [lmms-eval](https://github.com/EvolvingLMMs-Lab/lmms-eval).
121
+
122
+ #### - **LLaVA-NeXT: Stronger LLMs Supercharge Multimodal Capabilities in the Wild**
123
+ - [LLaVA-NeXT-Image](./docs/LLaVA-NeXT.md): for image demo inference and evaluation of stronger LMMs using [lmms-eval](https://github.com/EvolvingLMMs-Lab/lmms-eval).
124
+
125
+
126
+ #### - LLaVA-NeXT: A Strong Zero-shot Video Understanding Model
127
+ - [LLaVA-NeXT-Video](./docs/LLaVA-NeXT-Video.md): for video inference and evaluation scripts. We recommend to use [LMMs-video](https://lmms-lab.github.io/posts/lmms-eval-0.2/) for evaluation.
128
+
129
+ #### - LLaVA-NeXT: Tackling Multi-image, Video, and 3D in Large Multimodal Models
130
+ - [LLaVA-NeXT-Interleave](./docs/LLaVA-NeXT-Interleave.md): for multi-image demo and evaluation scripts.
131
+
132
+ ## SGLang for SpeedUp Inference and Deployment
133
+
134
+ We use [SGLang](https://github.com/sgl-project/sglang) to speed up inference and deployment of LLaVA-NeXT. You could make LLaVA-NeXT as a backend API service with SGLang.
135
+
136
+ **Prepare Environment**:
137
+ Following the instruction in the [sglang](https://github.com/sgl-project/sglang?tab=readme-ov-file#install)
138
+
139
+ ### LLaVA-NeXT/OneVision
140
+
141
+ Checkout the HTTP Post/Get and SRT usage at [sglang/examples/runtime/llava_onevision](https://github.com/sgl-project/sglang/tree/main/examples/runtime/llava_onevision)
142
+
143
+ ### LLaVA-NeXT (Video)
144
+
145
+ **Launch and Run on (K) Nodes**:
146
+ - Go to sglang project
147
+ ```
148
+ cd PATH_TO/sglang
149
+ ```
150
+ - First node:
151
+ ```sh
152
+ bash examples/usage/llava_video/srt_example_llava_v.sh K 0 YOUR_VIDEO_PATH YOUR_MODEL_PATH FRAMES_PER_VIDEO
153
+ (e.g. bash examples/usage/llava_video/srt_example_llava_v.sh K 0 examples/usage/llava_video/videos/Q98Z4OTh8RwmDonc.mp4 lmms-lab/LLaVA-NeXT-Video-7B-DPO 16)
154
+ ```
155
+ - Second node:
156
+ ```sh
157
+ bash examples/usage/llava_video/srt_example_llava_v.sh K 1 YOUR_VIDEO_PATH YOUR_MODEL_PATH FRAMES_PER_VIDEO
158
+ ```
159
+ - The K node:
160
+ ```sh
161
+ bash examples/usage/llava_video/srt_example_llava_v.sh K K-1 YOUR_VIDEO_PATH YOUR_MODEL_PATH FRAMES_PER_VIDEO
162
+ ```
163
+
164
+
165
+ ## Citation
166
+
167
+ If you find it useful for your research and applications, please cite related papers/blogs using this BibTeX:
168
+ ```bibtex
169
+ @article{li2024llava,
170
+ title={LLaVA-NeXT-Interleave: Tackling Multi-image, Video, and 3D in Large Multimodal Models},
171
+ author={Li, Feng and Zhang, Renrui and Zhang, Hao and Zhang, Yuanhan and Li, Bo and Li, Wei and Ma, Zejun and Li, Chunyuan},
172
+ journal={arXiv preprint arXiv:2407.07895},
173
+ year={2024}
174
+ }
175
+
176
+ @misc{li2024llavanext-ablations,
177
+ title={LLaVA-NeXT: What Else Influences Visual Instruction Tuning Beyond Data?},
178
+ url={https://llava-vl.github.io/blog/2024-05-25-llava-next-ablations/},
179
+ author={Li, Bo and Zhang, Hao and Zhang, Kaichen and Guo, Dong and Zhang, Yuanhan and Zhang, Renrui and Li, Feng and Liu, Ziwei and Li, Chunyuan},
180
+ month={May},
181
+ year={2024}
182
+ }
183
+
184
+ @misc{li2024llavanext-strong,
185
+ title={LLaVA-NeXT: Stronger LLMs Supercharge Multimodal Capabilities in the Wild},
186
+ url={https://llava-vl.github.io/blog/2024-05-10-llava-next-stronger-llms/},
187
+ author={Li, Bo and Zhang, Kaichen and Zhang, Hao and Guo, Dong and Zhang, Renrui and Li, Feng and Zhang, Yuanhan and Liu, Ziwei and Li, Chunyuan},
188
+ month={May},
189
+ year={2024}
190
+ }
191
+
192
+ @misc{zhang2024llavanext-video,
193
+ title={LLaVA-NeXT: A Strong Zero-shot Video Understanding Model},
194
+ url={https://llava-vl.github.io/blog/2024-04-30-llava-next-video/},
195
+ author={Zhang, Yuanhan and Li, Bo and Liu, haotian and Lee, Yong jae and Gui, Liangke and Fu, Di and Feng, Jiashi and Liu, Ziwei and Li, Chunyuan},
196
+ month={April},
197
+ year={2024}
198
+ }
199
+
200
+ @misc{liu2024llavanext,
201
+ title={LLaVA-NeXT: Improved reasoning, OCR, and world knowledge},
202
+ url={https://llava-vl.github.io/blog/2024-01-30-llava-next/},
203
+ author={Liu, Haotian and Li, Chunyuan and Li, Yuheng and Li, Bo and Zhang, Yuanhan and Shen, Sheng and Lee, Yong Jae},
204
+ month={January},
205
+ year={2024}
206
+ }
207
+
208
+ @misc{liu2023improvedllava,
209
+ title={Improved Baselines with Visual Instruction Tuning},
210
+ author={Liu, Haotian and Li, Chunyuan and Li, Yuheng and Lee, Yong Jae},
211
+ publisher={arXiv:2310.03744},
212
+ year={2023},
213
+ }
214
+
215
+ @misc{liu2023llava,
216
+ title={Visual Instruction Tuning},
217
+ author={Liu, Haotian and Li, Chunyuan and Wu, Qingyang and Lee, Yong Jae},
218
+ publisher={NeurIPS},
219
+ year={2023},
220
+ }
221
+ ```
222
+
223
+ ## Acknowledgement
224
+
225
+ - [Vicuna](https://github.com/lm-sys/FastChat): the codebase we built upon, and our base model Vicuna-13B that has the amazing language capabilities!
226
+ - The LLaVA-NeXT project is currently maintained by the team along with our contributors (listed alphabetically by the first names): [Bo Li](https://brianboli.com/), [Dong Guo](https://www.linkedin.com/in/dongguoset/), [Feng Li](https://scholar.google.com/citations?hl=zh-CN&user=ybRe9GcAAAAJ&view_op=list_works&sortby=pubdate), [Hao Zhang](https://scholar.google.com/citations?user=B8hPxMQAAAAJ&hl=en), [Kaichen Zhang](https://www.linkedin.com/in/kaichen-zhang-014b17219/?originalSubdomain=sg), [Renrui Zhang](https://zrrskywalker.github.io/), [Yuanhan Zhang](https://zhangyuanhan-ai.github.io/), led by [Chunyuan Li](https://chunyuan.li/) and with the guidance and help from [Haotian Liu](https://hliu.cc/).
227
+ - The `lmms-eval` framework and its core contributors, including Peiyuan Zhang, Fanyi Pu, Joshua Adrian Cahyono, and Kairui Hu, for their support on the evaluation side.
228
+
229
+ ## Related Projects
230
+
231
+ - [Instruction Tuning with GPT-4](https://github.com/Instruction-Tuning-with-GPT-4/GPT-4-LLM)
232
+ - [LLaVA-Med: Training a Large Language-and-Vision Assistant for Biomedicine in One Day](https://github.com/microsoft/LLaVA-Med)
233
+ - [Otter: In-Context Multi-Modal Instruction Tuning](https://github.com/Luodian/Otter)
234
+
235
+ For future project ideas, please check out:
236
+ - [SEEM: Segment Everything Everywhere All at Once](https://github.com/UX-Decoder/Segment-Everything-Everywhere-All-At-Once)
237
+ - [Grounded-Segment-Anything](https://github.com/IDEA-Research/Grounded-Segment-Anything) to detect, segment, and generate anything by marrying [Grounding DINO](https://github.com/IDEA-Research/GroundingDINO) and [Segment-Anything](https://github.com/facebookresearch/segment-anything).
all_layers_simulated_rope_qk_dot_product_corrected.pdf ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2c6c60314f1551fe06449dc2c97eadb0c921b1c769443df45819e02e69c184f
3
+ size 517033
all_layers_simulated_rope_qk_dot_product_corrected.png ADDED

Git LFS Details

  • SHA256: 1c0f2fcfb2756e62f9b970b20116f15509d949aa2d5b5290e4e98da1dc67a687
  • Pointer size: 131 Bytes
  • Size of remote file: 453 kB
all_layers_simulated_rope_qk_dot_product_corrected.svg ADDED
attention_heatmap.jpg ADDED
attention_map/mmpe/attention_heatmap_0.png ADDED
attention_map/mmpe/attention_heatmap_1.png ADDED
attention_map/mmpe/attention_heatmap_10.png ADDED
attention_map/mmpe/attention_heatmap_11.png ADDED
attention_map/mmpe/attention_heatmap_12.png ADDED
attention_map/mmpe/attention_heatmap_13.png ADDED
attention_map/mmpe/attention_heatmap_14.png ADDED
attention_map/mmpe/attention_heatmap_15.png ADDED
attention_map/mmpe/attention_heatmap_16.png ADDED
attention_map/mmpe/attention_heatmap_17.png ADDED
attention_map/mmpe/attention_heatmap_18.png ADDED
attention_map/mmpe/attention_heatmap_19.png ADDED
attention_map/mmpe/attention_heatmap_2.png ADDED
attention_map/mmpe/attention_heatmap_20.png ADDED
attention_map/mmpe/attention_heatmap_21.png ADDED
attention_map/mmpe/attention_heatmap_22.png ADDED
attention_map/mmpe/attention_heatmap_23.png ADDED
attention_map/mmpe/attention_heatmap_24.png ADDED
attention_map/mmpe/attention_heatmap_25.png ADDED
attention_map/mmpe/attention_heatmap_26.png ADDED
attention_map/mmpe/attention_heatmap_27.png ADDED
attention_map/mmpe/attention_heatmap_28.png ADDED
attention_map/mmpe/attention_heatmap_29.png ADDED
attention_map/mmpe/attention_heatmap_3.png ADDED
attention_map/mmpe/attention_heatmap_30.png ADDED
attention_map/mmpe/attention_heatmap_31.png ADDED
attention_map/mmpe/attention_heatmap_4.png ADDED
attention_map/mmpe/attention_heatmap_5.png ADDED
attention_map/mmpe/attention_heatmap_6.png ADDED
attention_map/mmpe/attention_heatmap_7.png ADDED
attention_map/mmpe/attention_heatmap_8.png ADDED
attention_map/mmpe/attention_heatmap_9.png ADDED
attention_map/mmpe/attn_layer0_330.png ADDED
attention_map/mmpe/attn_layer0_336.png ADDED
attention_map/mmpe/attn_layer0_948.png ADDED
attention_map/mmpe/attn_layer0_950.png ADDED
attention_map/mmpe/attn_layer0_952.png ADDED
attention_map/mmpe/attn_layer0_960.png ADDED
attention_map/mmpe/attn_layer0_between.png ADDED
attention_map/mmpe/attn_layer0_each_pair.png ADDED