Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +2 -0
- README.md +128 -3
- config.json +137 -0
- image1.jpg +0 -0
- internvl2_5_1b_364_ax630c/model.embed_tokens.weight.bfloat16.bin +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l0_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l10_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l11_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l12_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l13_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l14_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l15_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l16_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l17_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l18_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l19_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l1_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l20_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l21_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l22_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l23_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l2_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l3_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l4_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l5_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l6_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l7_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l8_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_p256_l9_together.axmodel +3 -0
- internvl2_5_1b_364_ax630c/qwen2_post.axmodel +3 -0
- internvl2_5_1b_364_ax630c/vit_intern_sim_space2depth.axmodel +3 -0
- internvl2_5_1b_448_ax650/internvl2_5_1b_mpo_vit.axmodel +3 -0
- internvl2_5_1b_448_ax650/model.embed_tokens.weight.bfloat16.bin +3 -0
- internvl2_5_1b_448_ax650/model.embed_tokens.weight.npy +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l0_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l10_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l11_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l12_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l13_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l14_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l15_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l16_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l17_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l18_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l19_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l1_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l20_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l21_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l22_together.axmodel +3 -0
- internvl2_5_1b_448_ax650/qwen2_p128_l23_together.axmodel +3 -0
.gitattributes
CHANGED
@@ -34,3 +34,5 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
*.axmodel filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
*.axmodel filter=lfs diff=lfs merge=lfs -text
|
37 |
+
main filter=lfs diff=lfs merge=lfs -text
|
38 |
+
main_ax650 filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -1,3 +1,128 @@
|
|
1 |
-
---
|
2 |
-
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
license: bsd-3-clause
|
4 |
+
base_model:
|
5 |
+
- OpenGVLab/InternVL2_5-1B
|
6 |
+
tags:
|
7 |
+
- InternVL2_5
|
8 |
+
- InternVL2_5-1B
|
9 |
+
- Int8
|
10 |
+
- VLM
|
11 |
+
---
|
12 |
+
|
13 |
+
# InternVL2_5-1B-Int8
|
14 |
+
|
15 |
+
This version of InternVL2_5-1B has been converted to run on the Axera NPU using **w8a16** quantization.
|
16 |
+
|
17 |
+
This model has been optimized with the following LoRA:
|
18 |
+
|
19 |
+
Compatible with Pulsar2 version: 3.3
|
20 |
+
|
21 |
+
## Convert tools links:
|
22 |
+
|
23 |
+
For those who are interested in model conversion, you can try to export axmodel through the original repo :
|
24 |
+
https://huggingface.co/OpenGVLab/InternVL2_5-1B-MPO
|
25 |
+
|
26 |
+
[Pulsar2 Link, How to Convert LLM from Huggingface to axmodel](https://pulsar2-docs.readthedocs.io/en/latest/appendix/build_llm.html)
|
27 |
+
|
28 |
+
[AXera NPU HOST LLM Runtime](https://github.com/AXERA-TECH/ax-llm/tree/internvl2)
|
29 |
+
|
30 |
+
[AXera NPU AXCL LLM Runtime](https://github.com/AXERA-TECH/ax-llm/tree/axcl-llm-internvl)
|
31 |
+
|
32 |
+
## Support Platform
|
33 |
+
|
34 |
+
- AX650
|
35 |
+
- AX650N DEMO Board
|
36 |
+
- [M4N-Dock(爱芯派Pro)](https://wiki.sipeed.com/hardware/zh/maixIV/m4ndock/m4ndock.html)
|
37 |
+
- [M.2 Accelerator card](https://axcl-docs.readthedocs.io/zh-cn/latest/doc_guide_hardware.html)
|
38 |
+
|
39 |
+
|Chips|image encoder 448|ttft|w8a16|
|
40 |
+
|--|--|--|--|
|
41 |
+
|AX650| 350 ms | 420 ms |32 tokens/sec|
|
42 |
+
|
43 |
+
- AX630C
|
44 |
+
- AX630C DEMO Board
|
45 |
+
- [M4N-Dock(爱芯派Pro)](https://wiki.sipeed.com/hardware/zh/maixIV/m4ndock/m4ndock.html)
|
46 |
+
- [M.2 Accelerator card](https://axcl-docs.readthedocs.io/zh-cn/latest/doc_guide_hardware.html)
|
47 |
+
- AX630C
|
48 |
+
|
49 |
+
|Chips|image encoder 364|ttft|w8a16|
|
50 |
+
|--|--|--|--|
|
51 |
+
|AX630C| 1769 ms | 1123 ms |2.4 tokens/sec|
|
52 |
+
|
53 |
+
## How to use
|
54 |
+
|
55 |
+
Download all files from this repository to the device
|
56 |
+
|
57 |
+
```
|
58 |
+
root@ax630c:InternVL2_5-1B-MPO-AX630C# tree -L 1
|
59 |
+
.
|
60 |
+
|-- config.json
|
61 |
+
|-- internvl2_5_1b_364_ax630c
|
62 |
+
|-- internvl2_5_tokenizer
|
63 |
+
|-- internvl2_5_tokenizer_364.py
|
64 |
+
|-- main
|
65 |
+
|-- run_internvl2_5_364_ax630c.sh
|
66 |
+
`-- image1.jpg
|
67 |
+
```
|
68 |
+
|
69 |
+
#### Install transformer
|
70 |
+
|
71 |
+
```
|
72 |
+
pip install transformers==4.41.1
|
73 |
+
```
|
74 |
+
|
75 |
+
#### Start the Tokenizer service
|
76 |
+
|
77 |
+
```
|
78 |
+
(vllm) lihongjie@gn4:InternVL2_5-1B-MPO-AX630C$ python internvl2_5_tokenizer_364.py --host localhost
|
79 |
+
None None 151645 <|im_end|>
|
80 |
+
[151644, 8948, 198, 56568, 104625, 100633, 104455, 104800, 101101, 32022, 102022, 99602, 100013, 9370, 90286, 21287, 42140, 53772, 35243, 26288, 104949, 3837, 105205, 109641, 67916, 30698, 11, 54851, 46944, 115404, 42192, 99441, 100623, 48692, 100168, 110498, 1773, 151645, 151644, 872, 198, 151665, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151667, 151666, 198, 5501, 7512, 279, 2168, 19620, 13, 151645, 151644, 77091, 198]
|
81 |
+
223
|
82 |
+
[151644, 8948, 198, 56568, 104625, 100633, 104455, 104800, 101101, 32022, 102022, 99602, 100013, 9370, 90286, 21287, 42140, 53772, 35243, 26288, 104949, 3837, 105205, 109641, 67916, 30698, 11, 54851, 46944, 115404, 42192, 99441, 100623, 48692, 100168, 110498, 1773, 151645, 151644, 872, 198, 14990, 1879, 151645, 151644, 77091, 198]
|
83 |
+
47
|
84 |
+
http://localhost:8080
|
85 |
+
```
|
86 |
+
|
87 |
+
#### Inference with AX630C Host, such as M4N-Dock(爱芯派Pro) or AX630C DEMO Board
|
88 |
+
|
89 |
+
- input text
|
90 |
+
|
91 |
+
```
|
92 |
+
Describe the picture
|
93 |
+
```
|
94 |
+
|
95 |
+
- input image
|
96 |
+
|
97 |
+

|
98 |
+
|
99 |
+
Open another terminal and run `./run_internvl2_5_364_ax630c.sh`
|
100 |
+
|
101 |
+
```
|
102 |
+
root@ax630c:InternVL2_5-1B-MPO-AX630C# bash run_internvl2_5_364_ax630c.sh
|
103 |
+
[I][ Init][ 106]: LLM init start
|
104 |
+
bos_id: -1, eos_id: 151645
|
105 |
+
3% | ██ | 1 / 28 [0.17s<4.90s, 5.71 count/s] tokenizer init ok[I][ Init][ 26]: LLaMaEmbedSelector use mmap
|
106 |
+
100% | ████████████████████████████████ | 28 / 28 [5.41s<5.41s, 5.17 count/s] init vpm axmodel ok,remain_cmm(907 MB)MB)
|
107 |
+
[I][ Init][ 254]: max_token_len : 1023
|
108 |
+
[I][ Init][ 259]: kv_cache_size : 128, kv_cache_num: 1023
|
109 |
+
[I][ Init][ 267]: prefill_token_num : 256
|
110 |
+
[I][ Init][ 269]: vpm_height : 364,vpm_width : 364
|
111 |
+
[I][ Init][ 278]: LLM init ok
|
112 |
+
Type "q" to exit, Ctrl+c to stop current running
|
113 |
+
prompt >> Please describe the image shortly.
|
114 |
+
image >> image1.jpg
|
115 |
+
[I][ Encode][ 337]: image encode time : 1768.706055 ms, size : 151424
|
116 |
+
[I][ Run][ 548]: ttft: 1123.02 ms
|
117 |
+
The image shows a red panda resting on a wooden platform. It has a reddish-brown fur coat with white markings around its eyes and ears. The background features green foliage.
|
118 |
+
|
119 |
+
[N][ Run][ 687]: hit eos,avg 2.42 token/s
|
120 |
+
|
121 |
+
prompt >>
|
122 |
+
```
|
123 |
+
|
124 |
+
#### Inference with M.2 Accelerator card
|
125 |
+
|
126 |
+
[What is M.2 Accelerator card?](https://axcl-docs.readthedocs.io/zh-cn/latest/doc_guide_hardware.html), Show this DEMO based on Raspberry PI 5.
|
127 |
+
|
128 |
+
*TODO*
|
config.json
ADDED
@@ -0,0 +1,137 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_commit_hash": null,
|
3 |
+
"architectures": [
|
4 |
+
"InternVLChatModel"
|
5 |
+
],
|
6 |
+
"auto_map": {
|
7 |
+
"AutoConfig": "configuration_internvl_chat.InternVLChatConfig",
|
8 |
+
"AutoModel": "modeling_internvl_chat.InternVLChatModel",
|
9 |
+
"AutoModelForCausalLM": "modeling_internvl_chat.InternVLChatModel"
|
10 |
+
},
|
11 |
+
"downsample_ratio": 0.5,
|
12 |
+
"dynamic_image_size": true,
|
13 |
+
"force_image_size": 448,
|
14 |
+
"llm_config": {
|
15 |
+
"_name_or_path": "Qwen/Qwen2.5-0.5B-Instruct",
|
16 |
+
"add_cross_attention": false,
|
17 |
+
"architectures": [
|
18 |
+
"Qwen2ForCausalLM"
|
19 |
+
],
|
20 |
+
"_attn_implementation": "flash_attention_2",
|
21 |
+
"attention_dropout": 0.0,
|
22 |
+
"bad_words_ids": null,
|
23 |
+
"begin_suppress_tokens": null,
|
24 |
+
"bos_token_id": 151643,
|
25 |
+
"chunk_size_feed_forward": 0,
|
26 |
+
"cross_attention_hidden_size": null,
|
27 |
+
"decoder_start_token_id": null,
|
28 |
+
"diversity_penalty": 0.0,
|
29 |
+
"do_sample": false,
|
30 |
+
"early_stopping": false,
|
31 |
+
"encoder_no_repeat_ngram_size": 0,
|
32 |
+
"eos_token_id": 151645,
|
33 |
+
"exponential_decay_length_penalty": null,
|
34 |
+
"finetuning_task": null,
|
35 |
+
"forced_bos_token_id": null,
|
36 |
+
"forced_eos_token_id": null,
|
37 |
+
"hidden_act": "silu",
|
38 |
+
"hidden_size": 896,
|
39 |
+
"id2label": {
|
40 |
+
"0": "LABEL_0",
|
41 |
+
"1": "LABEL_1"
|
42 |
+
},
|
43 |
+
"initializer_range": 0.02,
|
44 |
+
"intermediate_size": 4864,
|
45 |
+
"is_decoder": false,
|
46 |
+
"is_encoder_decoder": false,
|
47 |
+
"label2id": {
|
48 |
+
"LABEL_0": 0,
|
49 |
+
"LABEL_1": 1
|
50 |
+
},
|
51 |
+
"length_penalty": 1.0,
|
52 |
+
"max_length": 20,
|
53 |
+
"max_position_embeddings": 32768,
|
54 |
+
"max_window_layers": 21,
|
55 |
+
"min_length": 0,
|
56 |
+
"model_type": "qwen2",
|
57 |
+
"no_repeat_ngram_size": 0,
|
58 |
+
"num_attention_heads": 14,
|
59 |
+
"num_beam_groups": 1,
|
60 |
+
"num_beams": 1,
|
61 |
+
"num_hidden_layers": 24,
|
62 |
+
"num_key_value_heads": 2,
|
63 |
+
"num_return_sequences": 1,
|
64 |
+
"output_attentions": false,
|
65 |
+
"output_hidden_states": false,
|
66 |
+
"output_scores": false,
|
67 |
+
"pad_token_id": null,
|
68 |
+
"prefix": null,
|
69 |
+
"problem_type": null,
|
70 |
+
"pruned_heads": {},
|
71 |
+
"remove_invalid_values": false,
|
72 |
+
"repetition_penalty": 1.0,
|
73 |
+
"return_dict": true,
|
74 |
+
"return_dict_in_generate": false,
|
75 |
+
"rms_norm_eps": 1e-06,
|
76 |
+
"rope_theta": 1000000.0,
|
77 |
+
"sep_token_id": null,
|
78 |
+
"sliding_window": 32768,
|
79 |
+
"suppress_tokens": null,
|
80 |
+
"task_specific_params": null,
|
81 |
+
"temperature": 1.0,
|
82 |
+
"tf_legacy_loss": false,
|
83 |
+
"tie_encoder_decoder": false,
|
84 |
+
"tie_word_embeddings": false,
|
85 |
+
"tokenizer_class": null,
|
86 |
+
"top_k": 50,
|
87 |
+
"top_p": 1.0,
|
88 |
+
"torch_dtype": "bfloat16",
|
89 |
+
"torchscript": false,
|
90 |
+
"transformers_version": "4.37.2",
|
91 |
+
"typical_p": 1.0,
|
92 |
+
"use_bfloat16": true,
|
93 |
+
"use_cache": true,
|
94 |
+
"use_sliding_window": false,
|
95 |
+
"vocab_size": 151674
|
96 |
+
},
|
97 |
+
"max_dynamic_patch": 12,
|
98 |
+
"min_dynamic_patch": 1,
|
99 |
+
"model_type": "internvl_chat",
|
100 |
+
"ps_version": "v2",
|
101 |
+
"select_layer": -1,
|
102 |
+
"template": "internvl2_5",
|
103 |
+
"torch_dtype": "bfloat16",
|
104 |
+
"use_backbone_lora": 0,
|
105 |
+
"use_llm_lora": 0,
|
106 |
+
"use_thumbnail": true,
|
107 |
+
"vision_config": {
|
108 |
+
"architectures": [
|
109 |
+
"InternVisionModel"
|
110 |
+
],
|
111 |
+
"attention_dropout": 0.0,
|
112 |
+
"drop_path_rate": 0.0,
|
113 |
+
"dropout": 0.0,
|
114 |
+
"hidden_act": "gelu",
|
115 |
+
"hidden_size": 1024,
|
116 |
+
"image_size": 448,
|
117 |
+
"initializer_factor": 1.0,
|
118 |
+
"initializer_range": 0.02,
|
119 |
+
"intermediate_size": 4096,
|
120 |
+
"layer_norm_eps": 1e-06,
|
121 |
+
"model_type": "intern_vit_6b",
|
122 |
+
"norm_type": "layer_norm",
|
123 |
+
"num_attention_heads": 16,
|
124 |
+
"num_channels": 3,
|
125 |
+
"num_hidden_layers": 24,
|
126 |
+
"output_attentions": false,
|
127 |
+
"output_hidden_states": false,
|
128 |
+
"patch_size": 14,
|
129 |
+
"qk_normalization": false,
|
130 |
+
"qkv_bias": true,
|
131 |
+
"return_dict": true,
|
132 |
+
"torch_dtype": "bfloat16",
|
133 |
+
"transformers_version": "4.37.2",
|
134 |
+
"use_bfloat16": true,
|
135 |
+
"use_flash_attn": true
|
136 |
+
}
|
137 |
+
}
|
image1.jpg
ADDED
![]() |
internvl2_5_1b_364_ax630c/model.embed_tokens.weight.bfloat16.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c075733fbd78ce7da59e83318795095432ab287547e61ba21ad3fb05a4bb896b
|
3 |
+
size 271799808
|
internvl2_5_1b_364_ax630c/qwen2_p256_l0_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:af330da9bcebf41e4aea08de49d074954486d08d0b56131097d1aedbec7accb8
|
3 |
+
size 16009622
|
internvl2_5_1b_364_ax630c/qwen2_p256_l10_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5d672ce000c86706709e885e6aa586d299688a825fa9b2492743c068d0a3dcff
|
3 |
+
size 16009646
|
internvl2_5_1b_364_ax630c/qwen2_p256_l11_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:944652ed5157021334e990b9df3460a698853b2e4dbffd62255e701747467723
|
3 |
+
size 16009654
|
internvl2_5_1b_364_ax630c/qwen2_p256_l12_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fc2318e042d12e21108c0a7fd23e35978131da04408337319fa651cbb0ce8197
|
3 |
+
size 16009622
|
internvl2_5_1b_364_ax630c/qwen2_p256_l13_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:edaccef791b600fb84a94e1117b23b97b5ebd6f69a02875c3feef5f207740325
|
3 |
+
size 16009638
|
internvl2_5_1b_364_ax630c/qwen2_p256_l14_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8aba074ab790ce3a9048732c0c8606ba1e3a9606754d37f38a481762faa170ea
|
3 |
+
size 16009670
|
internvl2_5_1b_364_ax630c/qwen2_p256_l15_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:26d04515a1c97525d79e506865ddb527555bae3b4ce738519dff9693d63dd800
|
3 |
+
size 16009630
|
internvl2_5_1b_364_ax630c/qwen2_p256_l16_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b7e9c61713e674f37f3e1eb570085bda050aa7739718e4cb11f33a43a13d4250
|
3 |
+
size 16009630
|
internvl2_5_1b_364_ax630c/qwen2_p256_l17_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:682e958761bab38381aa38667adc6715f0b83e18062c884105ee92e135f0d5d1
|
3 |
+
size 16009630
|
internvl2_5_1b_364_ax630c/qwen2_p256_l18_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29defdfafd3ab57cca9e3602e6c92e3547788cba1ee9c2a627fd04f8ad07d3c6
|
3 |
+
size 16009638
|
internvl2_5_1b_364_ax630c/qwen2_p256_l19_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6897a6b66af8d9ef98bf1ea2aa810b9a2ef612f0b64ee47c8faec8e2b635122d
|
3 |
+
size 16009614
|
internvl2_5_1b_364_ax630c/qwen2_p256_l1_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a920d53223aa0a4746802b1705fef467763f371d09e0e1918d8e3b5c54311cb3
|
3 |
+
size 16009638
|
internvl2_5_1b_364_ax630c/qwen2_p256_l20_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:09f120a2b2f6619c90610c5027bf02f20ddadd12df5051e1c358f73c03cb2cf6
|
3 |
+
size 16009646
|
internvl2_5_1b_364_ax630c/qwen2_p256_l21_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d1209e8c33c9765d19bed54410aa83573aea3eab04754ff5d14885e401a2d5a2
|
3 |
+
size 16009630
|
internvl2_5_1b_364_ax630c/qwen2_p256_l22_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d47a40712c2d33989bd2e4c6889b458399edf4f72de9149942f484ebde0e972b
|
3 |
+
size 16009646
|
internvl2_5_1b_364_ax630c/qwen2_p256_l23_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:924a41dcf1618069182896f4a2e5e3058a8d72d26f5e537f6d7cf7162c2cc319
|
3 |
+
size 16009630
|
internvl2_5_1b_364_ax630c/qwen2_p256_l2_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9df37aad5c393dd516157551018b5c7bd1440d94959d6ee30229074bbf68a916
|
3 |
+
size 16009646
|
internvl2_5_1b_364_ax630c/qwen2_p256_l3_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e2d3efd9edcaf1eaeb613ddfd4f4d1568477e46426e20392d44e6b62190afa9e
|
3 |
+
size 16009622
|
internvl2_5_1b_364_ax630c/qwen2_p256_l4_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:16bef5efe6ddf2609db42aca42c5beb3bce6ccbecb5d4465a091c75e2e6ddc0d
|
3 |
+
size 16009646
|
internvl2_5_1b_364_ax630c/qwen2_p256_l5_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:df32fcff11d33fe96622734da6c9f11f05856c8494c1a853a8822efe6af9479e
|
3 |
+
size 16009614
|
internvl2_5_1b_364_ax630c/qwen2_p256_l6_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ccaf9ab7a2fd7e412e4992f8da48b4473ad417bf725df3332f1229ed5a68e6c1
|
3 |
+
size 16009630
|
internvl2_5_1b_364_ax630c/qwen2_p256_l7_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4cba84fcf8fd8cfe8f5b3bb2d097d2d423b80e9f8c2104b0eead26685a668a3d
|
3 |
+
size 16009638
|
internvl2_5_1b_364_ax630c/qwen2_p256_l8_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:08765f4dfbe87fe193b619285c289464fd16e9ac17b5f803af188c03726a5c2f
|
3 |
+
size 16009638
|
internvl2_5_1b_364_ax630c/qwen2_p256_l9_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:017870640a743a73853a903d8b03ef0b5998767e911b8f19a336cab3c4160cc2
|
3 |
+
size 16009638
|
internvl2_5_1b_364_ax630c/qwen2_post.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bb764b0fd3d62f22003eb3bc05c03a7250eb779d3cc1c440962c584c304b004c
|
3 |
+
size 139700408
|
internvl2_5_1b_364_ax630c/vit_intern_sim_space2depth.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c6a517a273eb945cb45922486db00cc884671dc38378f504c8d181de47fd024
|
3 |
+
size 402545119
|
internvl2_5_1b_448_ax650/internvl2_5_1b_mpo_vit.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:994f30ad7cad9db298436039dda8d2636865c7d569bab554ffb034df47ca964c
|
3 |
+
size 368051954
|
internvl2_5_1b_448_ax650/model.embed_tokens.weight.bfloat16.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9cecf2aa3a9d0762e525990561e75f08ab5df970d918f53e1b72dd73c81a6d34
|
3 |
+
size 271799808
|
internvl2_5_1b_448_ax650/model.embed_tokens.weight.npy
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0a0b4e9ae117fbec21ec4e3f0f1a0246e36ab155530ec3cfadc0fea905a00c15
|
3 |
+
size 543599744
|
internvl2_5_1b_448_ax650/qwen2_p128_l0_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1c4ed31c0eda91e7c5359b6db5c877c0884a17b71b10a0497c18a021d702cafa
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l10_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3b51816b79ea3c6ca7c1b464437ef7e616e64db1f18645d82eb53d80d92e671f
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l11_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b93695433fc6b0763a74257a03284b49810e351e6699f9dc5f36e848fd5eaaf6
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l12_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:221a39cc67d60630daddef139debfa48aeb7d0dd3c5b30b6fcc83fa3159ec01e
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l13_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6a21bb72ce4bfae62603bff502222eb8c8bbadd25902cd30510399428cdfa895
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l14_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:86f9807613275f91868cc30cee90cfe586c02f82425b8b081443fb35f6853e31
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l15_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ae8af7e2926600df551449da62daff088c10245f12f6a6a4d9637f9bbdc7817c
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l16_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f594ea1e9d024d302497f217fb885ceeb3947e69ced510946805fdd8ae2f082c
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l17_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d7141ef429665b1d16eff0802a337176b3f7474625c421cd2dc76102ffaf05db
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l18_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:043cc6a9d2df38f975c99770907327ba1a2c8bd584ac31aa09f944439a2b67e0
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l19_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8876f78bec716993e94463a41812e3b41eeb986dbb99db18f414e5e5e350110f
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l1_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0990b38144191ff3c84dcf221277882720a810f36e03b74dbac82d173ea996a7
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l20_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:29e201cc350b97ded7c00a15d0c0294342483ac657b249a0ebe2518d8bc0a77a
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l21_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c10872c59449bce9630a8cf861bb7855ec499d6968cf06594fac784dcc0822cf
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l22_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:147f52fba8790c089a39f374b06bb424efa92b9ef74b11951930b657a0d03022
|
3 |
+
size 20208870
|
internvl2_5_1b_448_ax650/qwen2_p128_l23_together.axmodel
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1b16e7ebd889d30e88ccb321737912c4665ad3dc8c610e34b66351b0b1d033bc
|
3 |
+
size 20208870
|