who-am-i commited on
Commit
622db33
·
verified ·
1 Parent(s): 86b384a

Upload folder using huggingface_hub

Browse files
Files changed (1) hide show
  1. config.json +7 -7
config.json CHANGED
@@ -1,8 +1,13 @@
1
  {
2
  "architectures": [
3
- "TmpModel"
4
  ],
5
  "attention_dropout": 0.0,
 
 
 
 
 
6
  "bos_token_id": 151643,
7
  "eos_token_id": 151645,
8
  "extra_feat_dim": 10,
@@ -103,10 +108,5 @@
103
  "vision_end_token_id": 151653,
104
  "vision_start_token_id": 151652,
105
  "vision_token_id": 151654,
106
- "vocab_size": 151936,
107
- "auto_map": {
108
- "AutoConfig": "configuration_qwen_rdx_linear.Qwen2_5_VLLinearConfig",
109
- "AutoModelForVision2Seq": "modeling_qwen_rdx_linear.Qwen2_5_VLLinearForCausalLM",
110
- "AutoModelForCausalLM": "modeling_qwen_rdx_linear.Qwen2_5_VLLinearForCausalLM"
111
- }
112
  }
 
1
  {
2
  "architectures": [
3
+ "Qwen2_5_VLLinearForCausalLM"
4
  ],
5
  "attention_dropout": 0.0,
6
+ "auto_map": {
7
+ "AutoConfig": "configuration_qwen_rdx_linear.Qwen2_5_VLLinearConfig",
8
+ "AutoModelForCausalLM": "modeling_qwen_rdx_linear.Qwen2_5_VLLinearForCausalLM",
9
+ "AutoModelForVision2Seq": "modeling_qwen_rdx_linear.Qwen2_5_VLLinearForCausalLM"
10
+ },
11
  "bos_token_id": 151643,
12
  "eos_token_id": 151645,
13
  "extra_feat_dim": 10,
 
108
  "vision_end_token_id": 151653,
109
  "vision_start_token_id": 151652,
110
  "vision_token_id": 151654,
111
+ "vocab_size": 151666
 
 
 
 
 
112
  }