qqc1989 commited on
Commit
3be80c4
·
verified ·
1 Parent(s): ada9866

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +2 -0
  2. README.md +128 -3
  3. config.json +137 -0
  4. image1.jpg +0 -0
  5. internvl2_5_1b_364_ax630c/model.embed_tokens.weight.bfloat16.bin +3 -0
  6. internvl2_5_1b_364_ax630c/qwen2_p256_l0_together.axmodel +3 -0
  7. internvl2_5_1b_364_ax630c/qwen2_p256_l10_together.axmodel +3 -0
  8. internvl2_5_1b_364_ax630c/qwen2_p256_l11_together.axmodel +3 -0
  9. internvl2_5_1b_364_ax630c/qwen2_p256_l12_together.axmodel +3 -0
  10. internvl2_5_1b_364_ax630c/qwen2_p256_l13_together.axmodel +3 -0
  11. internvl2_5_1b_364_ax630c/qwen2_p256_l14_together.axmodel +3 -0
  12. internvl2_5_1b_364_ax630c/qwen2_p256_l15_together.axmodel +3 -0
  13. internvl2_5_1b_364_ax630c/qwen2_p256_l16_together.axmodel +3 -0
  14. internvl2_5_1b_364_ax630c/qwen2_p256_l17_together.axmodel +3 -0
  15. internvl2_5_1b_364_ax630c/qwen2_p256_l18_together.axmodel +3 -0
  16. internvl2_5_1b_364_ax630c/qwen2_p256_l19_together.axmodel +3 -0
  17. internvl2_5_1b_364_ax630c/qwen2_p256_l1_together.axmodel +3 -0
  18. internvl2_5_1b_364_ax630c/qwen2_p256_l20_together.axmodel +3 -0
  19. internvl2_5_1b_364_ax630c/qwen2_p256_l21_together.axmodel +3 -0
  20. internvl2_5_1b_364_ax630c/qwen2_p256_l22_together.axmodel +3 -0
  21. internvl2_5_1b_364_ax630c/qwen2_p256_l23_together.axmodel +3 -0
  22. internvl2_5_1b_364_ax630c/qwen2_p256_l2_together.axmodel +3 -0
  23. internvl2_5_1b_364_ax630c/qwen2_p256_l3_together.axmodel +3 -0
  24. internvl2_5_1b_364_ax630c/qwen2_p256_l4_together.axmodel +3 -0
  25. internvl2_5_1b_364_ax630c/qwen2_p256_l5_together.axmodel +3 -0
  26. internvl2_5_1b_364_ax630c/qwen2_p256_l6_together.axmodel +3 -0
  27. internvl2_5_1b_364_ax630c/qwen2_p256_l7_together.axmodel +3 -0
  28. internvl2_5_1b_364_ax630c/qwen2_p256_l8_together.axmodel +3 -0
  29. internvl2_5_1b_364_ax630c/qwen2_p256_l9_together.axmodel +3 -0
  30. internvl2_5_1b_364_ax630c/qwen2_post.axmodel +3 -0
  31. internvl2_5_1b_364_ax630c/vit_intern_sim_space2depth.axmodel +3 -0
  32. internvl2_5_1b_448_ax650/internvl2_5_1b_mpo_vit.axmodel +3 -0
  33. internvl2_5_1b_448_ax650/model.embed_tokens.weight.bfloat16.bin +3 -0
  34. internvl2_5_1b_448_ax650/model.embed_tokens.weight.npy +3 -0
  35. internvl2_5_1b_448_ax650/qwen2_p128_l0_together.axmodel +3 -0
  36. internvl2_5_1b_448_ax650/qwen2_p128_l10_together.axmodel +3 -0
  37. internvl2_5_1b_448_ax650/qwen2_p128_l11_together.axmodel +3 -0
  38. internvl2_5_1b_448_ax650/qwen2_p128_l12_together.axmodel +3 -0
  39. internvl2_5_1b_448_ax650/qwen2_p128_l13_together.axmodel +3 -0
  40. internvl2_5_1b_448_ax650/qwen2_p128_l14_together.axmodel +3 -0
  41. internvl2_5_1b_448_ax650/qwen2_p128_l15_together.axmodel +3 -0
  42. internvl2_5_1b_448_ax650/qwen2_p128_l16_together.axmodel +3 -0
  43. internvl2_5_1b_448_ax650/qwen2_p128_l17_together.axmodel +3 -0
  44. internvl2_5_1b_448_ax650/qwen2_p128_l18_together.axmodel +3 -0
  45. internvl2_5_1b_448_ax650/qwen2_p128_l19_together.axmodel +3 -0
  46. internvl2_5_1b_448_ax650/qwen2_p128_l1_together.axmodel +3 -0
  47. internvl2_5_1b_448_ax650/qwen2_p128_l20_together.axmodel +3 -0
  48. internvl2_5_1b_448_ax650/qwen2_p128_l21_together.axmodel +3 -0
  49. internvl2_5_1b_448_ax650/qwen2_p128_l22_together.axmodel +3 -0
  50. internvl2_5_1b_448_ax650/qwen2_p128_l23_together.axmodel +3 -0
.gitattributes CHANGED
@@ -34,3 +34,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  *.axmodel filter=lfs diff=lfs merge=lfs -text
 
 
 
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
  *.axmodel filter=lfs diff=lfs merge=lfs -text
37
+ main filter=lfs diff=lfs merge=lfs -text
38
+ main_ax650 filter=lfs diff=lfs merge=lfs -text
README.md CHANGED
@@ -1,3 +1,128 @@
1
- ---
2
- license: mit
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ license: bsd-3-clause
4
+ base_model:
5
+ - OpenGVLab/InternVL2_5-1B
6
+ tags:
7
+ - InternVL2_5
8
+ - InternVL2_5-1B
9
+ - Int8
10
+ - VLM
11
+ ---
12
+
13
+ # InternVL2_5-1B-Int8
14
+
15
+ This version of InternVL2_5-1B has been converted to run on the Axera NPU using **w8a16** quantization.
16
+
17
+ This model has been optimized with the following LoRA:
18
+
19
+ Compatible with Pulsar2 version: 3.3
20
+
21
+ ## Convert tools links:
22
+
23
+ For those who are interested in model conversion, you can try to export axmodel through the original repo :
24
+ https://huggingface.co/OpenGVLab/InternVL2_5-1B-MPO
25
+
26
+ [Pulsar2 Link, How to Convert LLM from Huggingface to axmodel](https://pulsar2-docs.readthedocs.io/en/latest/appendix/build_llm.html)
27
+
28
+ [AXera NPU HOST LLM Runtime](https://github.com/AXERA-TECH/ax-llm/tree/internvl2)
29
+
30
+ [AXera NPU AXCL LLM Runtime](https://github.com/AXERA-TECH/ax-llm/tree/axcl-llm-internvl)
31
+
32
+ ## Support Platform
33
+
34
+ - AX650
35
+ - AX650N DEMO Board
36
+ - [M4N-Dock(爱芯派Pro)](https://wiki.sipeed.com/hardware/zh/maixIV/m4ndock/m4ndock.html)
37
+ - [M.2 Accelerator card](https://axcl-docs.readthedocs.io/zh-cn/latest/doc_guide_hardware.html)
38
+
39
+ |Chips|image encoder 448|ttft|w8a16|
40
+ |--|--|--|--|
41
+ |AX650| 350 ms | 420 ms |32 tokens/sec|
42
+
43
+ - AX630C
44
+ - AX630C DEMO Board
45
+ - [M4N-Dock(爱芯派Pro)](https://wiki.sipeed.com/hardware/zh/maixIV/m4ndock/m4ndock.html)
46
+ - [M.2 Accelerator card](https://axcl-docs.readthedocs.io/zh-cn/latest/doc_guide_hardware.html)
47
+ - AX630C
48
+
49
+ |Chips|image encoder 364|ttft|w8a16|
50
+ |--|--|--|--|
51
+ |AX630C| 1769 ms | 1123 ms |2.4 tokens/sec|
52
+
53
+ ## How to use
54
+
55
+ Download all files from this repository to the device
56
+
57
+ ```
58
+ root@ax630c:InternVL2_5-1B-MPO-AX630C# tree -L 1
59
+ .
60
+ |-- config.json
61
+ |-- internvl2_5_1b_364_ax630c
62
+ |-- internvl2_5_tokenizer
63
+ |-- internvl2_5_tokenizer_364.py
64
+ |-- main
65
+ |-- run_internvl2_5_364_ax630c.sh
66
+ `-- image1.jpg
67
+ ```
68
+
69
+ #### Install transformer
70
+
71
+ ```
72
+ pip install transformers==4.41.1
73
+ ```
74
+
75
+ #### Start the Tokenizer service
76
+
77
+ ```
78
+ (vllm) lihongjie@gn4:InternVL2_5-1B-MPO-AX630C$ python internvl2_5_tokenizer_364.py --host localhost
79
+ None None 151645 <|im_end|>
80
+ [151644, 8948, 198, 56568, 104625, 100633, 104455, 104800, 101101, 32022, 102022, 99602, 100013, 9370, 90286, 21287, 42140, 53772, 35243, 26288, 104949, 3837, 105205, 109641, 67916, 30698, 11, 54851, 46944, 115404, 42192, 99441, 100623, 48692, 100168, 110498, 1773, 151645, 151644, 872, 198, 151665, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151666, 198, 5501, 7512, 279, 2168, 19620, 13, 151645, 151644, 77091, 198]
81
+ 223
82
+ [151644, 8948, 198, 56568, 104625, 100633, 104455, 104800, 101101, 32022, 102022, 99602, 100013, 9370, 90286, 21287, 42140, 53772, 35243, 26288, 104949, 3837, 105205, 109641, 67916, 30698, 11, 54851, 46944, 115404, 42192, 99441, 100623, 48692, 100168, 110498, 1773, 151645, 151644, 872, 198, 14990, 1879, 151645, 151644, 77091, 198]
83
+ 47
84
+ http://localhost:8080
85
+ ```
86
+
87
+ #### Inference with AX630C Host, such as M4N-Dock(爱芯派Pro) or AX630C DEMO Board
88
+
89
+ - input text
90
+
91
+ ```
92
+ Describe the picture
93
+ ```
94
+
95
+ - input image
96
+
97
+ ![](./ssd_car.jpg)
98
+
99
+ Open another terminal and run `./run_internvl2_5_364_ax630c.sh`
100
+
101
+ ```
102
+ root@ax630c:InternVL2_5-1B-MPO-AX630C# bash run_internvl2_5_364_ax630c.sh
103
+ [I][ Init][ 106]: LLM init start
104
+ bos_id: -1, eos_id: 151645
105
+ 3% | ██ | 1 / 28 [0.17s<4.90s, 5.71 count/s] tokenizer init ok[I][ Init][ 26]: LLaMaEmbedSelector use mmap
106
+ 100% | ████████████████████████████████ | 28 / 28 [5.41s<5.41s, 5.17 count/s] init vpm axmodel ok,remain_cmm(907 MB)MB)
107
+ [I][ Init][ 254]: max_token_len : 1023
108
+ [I][ Init][ 259]: kv_cache_size : 128, kv_cache_num: 1023
109
+ [I][ Init][ 267]: prefill_token_num : 256
110
+ [I][ Init][ 269]: vpm_height : 364,vpm_width : 364
111
+ [I][ Init][ 278]: LLM init ok
112
+ Type "q" to exit, Ctrl+c to stop current running
113
+ prompt >> Please describe the image shortly.
114
+ image >> image1.jpg
115
+ [I][ Encode][ 337]: image encode time : 1768.706055 ms, size : 151424
116
+ [I][ Run][ 548]: ttft: 1123.02 ms
117
+ The image shows a red panda resting on a wooden platform. It has a reddish-brown fur coat with white markings around its eyes and ears. The background features green foliage.
118
+
119
+ [N][ Run][ 687]: hit eos,avg 2.42 token/s
120
+
121
+ prompt >>
122
+ ```
123
+
124
+ #### Inference with M.2 Accelerator card
125
+
126
+ [What is M.2 Accelerator card?](https://axcl-docs.readthedocs.io/zh-cn/latest/doc_guide_hardware.html), Show this DEMO based on Raspberry PI 5.
127
+
128
+ *TODO*
config.json ADDED
@@ -0,0 +1,137 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_commit_hash": null,
3
+ "architectures": [
4
+ "InternVLChatModel"
5
+ ],
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_internvl_chat.InternVLChatConfig",
8
+ "AutoModel": "modeling_internvl_chat.InternVLChatModel",
9
+ "AutoModelForCausalLM": "modeling_internvl_chat.InternVLChatModel"
10
+ },
11
+ "downsample_ratio": 0.5,
12
+ "dynamic_image_size": true,
13
+ "force_image_size": 448,
14
+ "llm_config": {
15
+ "_name_or_path": "Qwen/Qwen2.5-0.5B-Instruct",
16
+ "add_cross_attention": false,
17
+ "architectures": [
18
+ "Qwen2ForCausalLM"
19
+ ],
20
+ "_attn_implementation": "flash_attention_2",
21
+ "attention_dropout": 0.0,
22
+ "bad_words_ids": null,
23
+ "begin_suppress_tokens": null,
24
+ "bos_token_id": 151643,
25
+ "chunk_size_feed_forward": 0,
26
+ "cross_attention_hidden_size": null,
27
+ "decoder_start_token_id": null,
28
+ "diversity_penalty": 0.0,
29
+ "do_sample": false,
30
+ "early_stopping": false,
31
+ "encoder_no_repeat_ngram_size": 0,
32
+ "eos_token_id": 151645,
33
+ "exponential_decay_length_penalty": null,
34
+ "finetuning_task": null,
35
+ "forced_bos_token_id": null,
36
+ "forced_eos_token_id": null,
37
+ "hidden_act": "silu",
38
+ "hidden_size": 896,
39
+ "id2label": {
40
+ "0": "LABEL_0",
41
+ "1": "LABEL_1"
42
+ },
43
+ "initializer_range": 0.02,
44
+ "intermediate_size": 4864,
45
+ "is_decoder": false,
46
+ "is_encoder_decoder": false,
47
+ "label2id": {
48
+ "LABEL_0": 0,
49
+ "LABEL_1": 1
50
+ },
51
+ "length_penalty": 1.0,
52
+ "max_length": 20,
53
+ "max_position_embeddings": 32768,
54
+ "max_window_layers": 21,
55
+ "min_length": 0,
56
+ "model_type": "qwen2",
57
+ "no_repeat_ngram_size": 0,
58
+ "num_attention_heads": 14,
59
+ "num_beam_groups": 1,
60
+ "num_beams": 1,
61
+ "num_hidden_layers": 24,
62
+ "num_key_value_heads": 2,
63
+ "num_return_sequences": 1,
64
+ "output_attentions": false,
65
+ "output_hidden_states": false,
66
+ "output_scores": false,
67
+ "pad_token_id": null,
68
+ "prefix": null,
69
+ "problem_type": null,
70
+ "pruned_heads": {},
71
+ "remove_invalid_values": false,
72
+ "repetition_penalty": 1.0,
73
+ "return_dict": true,
74
+ "return_dict_in_generate": false,
75
+ "rms_norm_eps": 1e-06,
76
+ "rope_theta": 1000000.0,
77
+ "sep_token_id": null,
78
+ "sliding_window": 32768,
79
+ "suppress_tokens": null,
80
+ "task_specific_params": null,
81
+ "temperature": 1.0,
82
+ "tf_legacy_loss": false,
83
+ "tie_encoder_decoder": false,
84
+ "tie_word_embeddings": false,
85
+ "tokenizer_class": null,
86
+ "top_k": 50,
87
+ "top_p": 1.0,
88
+ "torch_dtype": "bfloat16",
89
+ "torchscript": false,
90
+ "transformers_version": "4.37.2",
91
+ "typical_p": 1.0,
92
+ "use_bfloat16": true,
93
+ "use_cache": true,
94
+ "use_sliding_window": false,
95
+ "vocab_size": 151674
96
+ },
97
+ "max_dynamic_patch": 12,
98
+ "min_dynamic_patch": 1,
99
+ "model_type": "internvl_chat",
100
+ "ps_version": "v2",
101
+ "select_layer": -1,
102
+ "template": "internvl2_5",
103
+ "torch_dtype": "bfloat16",
104
+ "use_backbone_lora": 0,
105
+ "use_llm_lora": 0,
106
+ "use_thumbnail": true,
107
+ "vision_config": {
108
+ "architectures": [
109
+ "InternVisionModel"
110
+ ],
111
+ "attention_dropout": 0.0,
112
+ "drop_path_rate": 0.0,
113
+ "dropout": 0.0,
114
+ "hidden_act": "gelu",
115
+ "hidden_size": 1024,
116
+ "image_size": 448,
117
+ "initializer_factor": 1.0,
118
+ "initializer_range": 0.02,
119
+ "intermediate_size": 4096,
120
+ "layer_norm_eps": 1e-06,
121
+ "model_type": "intern_vit_6b",
122
+ "norm_type": "layer_norm",
123
+ "num_attention_heads": 16,
124
+ "num_channels": 3,
125
+ "num_hidden_layers": 24,
126
+ "output_attentions": false,
127
+ "output_hidden_states": false,
128
+ "patch_size": 14,
129
+ "qk_normalization": false,
130
+ "qkv_bias": true,
131
+ "return_dict": true,
132
+ "torch_dtype": "bfloat16",
133
+ "transformers_version": "4.37.2",
134
+ "use_bfloat16": true,
135
+ "use_flash_attn": true
136
+ }
137
+ }
image1.jpg ADDED
internvl2_5_1b_364_ax630c/model.embed_tokens.weight.bfloat16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c075733fbd78ce7da59e83318795095432ab287547e61ba21ad3fb05a4bb896b
3
+ size 271799808
internvl2_5_1b_364_ax630c/qwen2_p256_l0_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:af330da9bcebf41e4aea08de49d074954486d08d0b56131097d1aedbec7accb8
3
+ size 16009622
internvl2_5_1b_364_ax630c/qwen2_p256_l10_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d672ce000c86706709e885e6aa586d299688a825fa9b2492743c068d0a3dcff
3
+ size 16009646
internvl2_5_1b_364_ax630c/qwen2_p256_l11_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:944652ed5157021334e990b9df3460a698853b2e4dbffd62255e701747467723
3
+ size 16009654
internvl2_5_1b_364_ax630c/qwen2_p256_l12_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fc2318e042d12e21108c0a7fd23e35978131da04408337319fa651cbb0ce8197
3
+ size 16009622
internvl2_5_1b_364_ax630c/qwen2_p256_l13_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:edaccef791b600fb84a94e1117b23b97b5ebd6f69a02875c3feef5f207740325
3
+ size 16009638
internvl2_5_1b_364_ax630c/qwen2_p256_l14_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8aba074ab790ce3a9048732c0c8606ba1e3a9606754d37f38a481762faa170ea
3
+ size 16009670
internvl2_5_1b_364_ax630c/qwen2_p256_l15_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26d04515a1c97525d79e506865ddb527555bae3b4ce738519dff9693d63dd800
3
+ size 16009630
internvl2_5_1b_364_ax630c/qwen2_p256_l16_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7e9c61713e674f37f3e1eb570085bda050aa7739718e4cb11f33a43a13d4250
3
+ size 16009630
internvl2_5_1b_364_ax630c/qwen2_p256_l17_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:682e958761bab38381aa38667adc6715f0b83e18062c884105ee92e135f0d5d1
3
+ size 16009630
internvl2_5_1b_364_ax630c/qwen2_p256_l18_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29defdfafd3ab57cca9e3602e6c92e3547788cba1ee9c2a627fd04f8ad07d3c6
3
+ size 16009638
internvl2_5_1b_364_ax630c/qwen2_p256_l19_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6897a6b66af8d9ef98bf1ea2aa810b9a2ef612f0b64ee47c8faec8e2b635122d
3
+ size 16009614
internvl2_5_1b_364_ax630c/qwen2_p256_l1_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a920d53223aa0a4746802b1705fef467763f371d09e0e1918d8e3b5c54311cb3
3
+ size 16009638
internvl2_5_1b_364_ax630c/qwen2_p256_l20_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09f120a2b2f6619c90610c5027bf02f20ddadd12df5051e1c358f73c03cb2cf6
3
+ size 16009646
internvl2_5_1b_364_ax630c/qwen2_p256_l21_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d1209e8c33c9765d19bed54410aa83573aea3eab04754ff5d14885e401a2d5a2
3
+ size 16009630
internvl2_5_1b_364_ax630c/qwen2_p256_l22_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d47a40712c2d33989bd2e4c6889b458399edf4f72de9149942f484ebde0e972b
3
+ size 16009646
internvl2_5_1b_364_ax630c/qwen2_p256_l23_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:924a41dcf1618069182896f4a2e5e3058a8d72d26f5e537f6d7cf7162c2cc319
3
+ size 16009630
internvl2_5_1b_364_ax630c/qwen2_p256_l2_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9df37aad5c393dd516157551018b5c7bd1440d94959d6ee30229074bbf68a916
3
+ size 16009646
internvl2_5_1b_364_ax630c/qwen2_p256_l3_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e2d3efd9edcaf1eaeb613ddfd4f4d1568477e46426e20392d44e6b62190afa9e
3
+ size 16009622
internvl2_5_1b_364_ax630c/qwen2_p256_l4_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16bef5efe6ddf2609db42aca42c5beb3bce6ccbecb5d4465a091c75e2e6ddc0d
3
+ size 16009646
internvl2_5_1b_364_ax630c/qwen2_p256_l5_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df32fcff11d33fe96622734da6c9f11f05856c8494c1a853a8822efe6af9479e
3
+ size 16009614
internvl2_5_1b_364_ax630c/qwen2_p256_l6_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccaf9ab7a2fd7e412e4992f8da48b4473ad417bf725df3332f1229ed5a68e6c1
3
+ size 16009630
internvl2_5_1b_364_ax630c/qwen2_p256_l7_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4cba84fcf8fd8cfe8f5b3bb2d097d2d423b80e9f8c2104b0eead26685a668a3d
3
+ size 16009638
internvl2_5_1b_364_ax630c/qwen2_p256_l8_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:08765f4dfbe87fe193b619285c289464fd16e9ac17b5f803af188c03726a5c2f
3
+ size 16009638
internvl2_5_1b_364_ax630c/qwen2_p256_l9_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:017870640a743a73853a903d8b03ef0b5998767e911b8f19a336cab3c4160cc2
3
+ size 16009638
internvl2_5_1b_364_ax630c/qwen2_post.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb764b0fd3d62f22003eb3bc05c03a7250eb779d3cc1c440962c584c304b004c
3
+ size 139700408
internvl2_5_1b_364_ax630c/vit_intern_sim_space2depth.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c6a517a273eb945cb45922486db00cc884671dc38378f504c8d181de47fd024
3
+ size 402545119
internvl2_5_1b_448_ax650/internvl2_5_1b_mpo_vit.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:994f30ad7cad9db298436039dda8d2636865c7d569bab554ffb034df47ca964c
3
+ size 368051954
internvl2_5_1b_448_ax650/model.embed_tokens.weight.bfloat16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9cecf2aa3a9d0762e525990561e75f08ab5df970d918f53e1b72dd73c81a6d34
3
+ size 271799808
internvl2_5_1b_448_ax650/model.embed_tokens.weight.npy ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0a0b4e9ae117fbec21ec4e3f0f1a0246e36ab155530ec3cfadc0fea905a00c15
3
+ size 543599744
internvl2_5_1b_448_ax650/qwen2_p128_l0_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1c4ed31c0eda91e7c5359b6db5c877c0884a17b71b10a0497c18a021d702cafa
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l10_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b51816b79ea3c6ca7c1b464437ef7e616e64db1f18645d82eb53d80d92e671f
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l11_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b93695433fc6b0763a74257a03284b49810e351e6699f9dc5f36e848fd5eaaf6
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l12_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:221a39cc67d60630daddef139debfa48aeb7d0dd3c5b30b6fcc83fa3159ec01e
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l13_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6a21bb72ce4bfae62603bff502222eb8c8bbadd25902cd30510399428cdfa895
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l14_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:86f9807613275f91868cc30cee90cfe586c02f82425b8b081443fb35f6853e31
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l15_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ae8af7e2926600df551449da62daff088c10245f12f6a6a4d9637f9bbdc7817c
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l16_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f594ea1e9d024d302497f217fb885ceeb3947e69ced510946805fdd8ae2f082c
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l17_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d7141ef429665b1d16eff0802a337176b3f7474625c421cd2dc76102ffaf05db
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l18_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:043cc6a9d2df38f975c99770907327ba1a2c8bd584ac31aa09f944439a2b67e0
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l19_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8876f78bec716993e94463a41812e3b41eeb986dbb99db18f414e5e5e350110f
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l1_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0990b38144191ff3c84dcf221277882720a810f36e03b74dbac82d173ea996a7
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l20_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:29e201cc350b97ded7c00a15d0c0294342483ac657b249a0ebe2518d8bc0a77a
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l21_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c10872c59449bce9630a8cf861bb7855ec499d6968cf06594fac784dcc0822cf
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l22_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:147f52fba8790c089a39f374b06bb424efa92b9ef74b11951930b657a0d03022
3
+ size 20208870
internvl2_5_1b_448_ax650/qwen2_p128_l23_together.axmodel ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1b16e7ebd889d30e88ccb321737912c4665ad3dc8c610e34b66351b0b1d033bc
3
+ size 20208870