khuam commited on
Commit
a17adef
·
verified ·
1 Parent(s): ab0d854

Model save

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: google/gemma-3-4b-it
3
  library_name: transformers
4
  model_name: run_1
5
  tags:
@@ -11,7 +11,7 @@ licence: license
11
 
12
  # Model Card for run_1
13
 
14
- This model is a fine-tuned version of [google/gemma-3-4b-it](https://huggingface.co/google/gemma-3-4b-it).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
 
1
  ---
2
+ base_model: Qwen/Qwen2.5-VL-3B-Instruct
3
  library_name: transformers
4
  model_name: run_1
5
  tags:
 
11
 
12
  # Model Card for run_1
13
 
14
+ This model is a fine-tuned version of [Qwen/Qwen2.5-VL-3B-Instruct](https://huggingface.co/Qwen/Qwen2.5-VL-3B-Instruct).
15
  It has been trained using [TRL](https://github.com/huggingface/trl).
16
 
17
  ## Quick start
adapter_checkpoint/README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: google/gemma-3-4b-it
3
  library_name: peft
4
  ---
5
 
 
1
  ---
2
+ base_model: Qwen/Qwen2.5-VL-3B-Instruct
3
  library_name: peft
4
  ---
5
 
adapter_checkpoint/adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "google/gemma-3-4b-it",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
@@ -27,16 +27,17 @@
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
30
- "fc1",
31
- "o_proj",
32
- "down_proj",
33
- "q_proj",
34
- "out_proj",
35
  "k_proj",
36
- "v_proj",
37
- "fc2",
38
  "up_proj",
39
- "gate_proj"
 
 
 
 
 
 
 
40
  ],
41
  "task_type": "CAUSAL_LM",
42
  "trainable_token_indices": null,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "Qwen/Qwen2.5-VL-3B-Instruct",
5
  "bias": "none",
6
  "corda_config": null,
7
  "eva_config": null,
 
27
  "rank_pattern": {},
28
  "revision": null,
29
  "target_modules": [
 
 
 
 
 
30
  "k_proj",
31
+ "mlp.0",
 
32
  "up_proj",
33
+ "gate_proj",
34
+ "mlp.2",
35
+ "v_proj",
36
+ "attn.proj",
37
+ "down_proj",
38
+ "o_proj",
39
+ "qkv",
40
+ "q_proj"
41
  ],
42
  "task_type": "CAUSAL_LM",
43
  "trainable_token_indices": null,
adapter_checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b246b6307a51e02f7023721b108aa369aff1e3b07515b3e8aee41f5253d5b8e9
3
- size 2762127280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7c10a696ea20bab93f1263580f24a4adc032b39634812062604adafa9b9016ca
3
+ size 1327495272
adapter_checkpoint/added_tokens.json CHANGED
@@ -1,3 +1,24 @@
1
  {
2
- "<image_soft_token>": 262144
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  }
 
1
  {
2
+ "</tool_call>": 151658,
3
+ "<tool_call>": 151657,
4
+ "<|box_end|>": 151649,
5
+ "<|box_start|>": 151648,
6
+ "<|endoftext|>": 151643,
7
+ "<|file_sep|>": 151664,
8
+ "<|fim_middle|>": 151660,
9
+ "<|fim_pad|>": 151662,
10
+ "<|fim_prefix|>": 151659,
11
+ "<|fim_suffix|>": 151661,
12
+ "<|im_end|>": 151645,
13
+ "<|im_start|>": 151644,
14
+ "<|image_pad|>": 151655,
15
+ "<|object_ref_end|>": 151647,
16
+ "<|object_ref_start|>": 151646,
17
+ "<|quad_end|>": 151651,
18
+ "<|quad_start|>": 151650,
19
+ "<|repo_name|>": 151663,
20
+ "<|video_pad|>": 151656,
21
+ "<|vision_end|>": 151653,
22
+ "<|vision_pad|>": 151654,
23
+ "<|vision_start|>": 151652
24
  }
adapter_checkpoint/chat_template.json CHANGED
@@ -1,3 +1,3 @@
1
  {
2
- "chat_template": "{{ bos_token }}\n{%- if messages[0]['role'] == 'system' -%}\n {%- if messages[0]['content'] is string -%}\n {%- set first_user_prefix = messages[0]['content'] + '\n\n' -%}\n {%- else -%}\n {%- set first_user_prefix = messages[0]['content'][0]['text'] + '\n\n' -%}\n {%- endif -%}\n {%- set loop_messages = messages[1:] -%}\n{%- else -%}\n {%- set first_user_prefix = \"\" -%}\n {%- set loop_messages = messages -%}\n{%- endif -%}\n{%- for message in loop_messages -%}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) -%}\n {{ raise_exception(\"Conversation roles must alternate user/assistant/user/assistant/...\") }}\n {%- endif -%}\n {%- if (message['role'] == 'assistant') -%}\n {%- set role = \"model\" -%}\n {%- else -%}\n {%- set role = message['role'] -%}\n {%- endif -%}\n {{ '<start_of_turn>' + role + '\n' + (first_user_prefix if loop.first else \"\") }}\n {%- if message['content'] is string -%}\n {{ message['content'] | trim }}\n {%- elif message['content'] is iterable -%}\n {%- for item in message['content'] -%}\n {%- if item['type'] == 'image' -%}\n {{ '<start_of_image>' }}\n {%- elif item['type'] == 'text' -%}\n {{ item['text'] | trim }}\n {%- endif -%}\n {%- endfor -%}\n {%- else -%}\n {{ raise_exception(\"Invalid content type\") }}\n {%- endif -%}\n {{ '<end_of_turn>\n' }}\n{%- endfor -%}\n{%- if add_generation_prompt -%}\n {{'<start_of_turn>model\n'}}\n{%- endif -%}\n"
3
  }
 
1
  {
2
+ "chat_template": "{% set image_count = namespace(value=0) %}{% set video_count = namespace(value=0) %}{% for message in messages %}{% if loop.first and message['role'] != 'system' %}<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n{% endif %}<|im_start|>{{ message['role'] }}\n{% if message['content'] is string %}{{ message['content'] }}<|im_end|>\n{% else %}{% for content in message['content'] %}{% if content['type'] == 'image' or 'image' in content or 'image_url' in content %}{% set image_count.value = image_count.value + 1 %}{% if add_vision_id %}Picture {{ image_count.value }}: {% endif %}<|vision_start|><|image_pad|><|vision_end|>{% elif content['type'] == 'video' or 'video' in content %}{% set video_count.value = video_count.value + 1 %}{% if add_vision_id %}Video {{ video_count.value }}: {% endif %}<|vision_start|><|video_pad|><|vision_end|>{% elif 'text' in content %}{{ content['text'] }}{% endif %}{% endfor %}<|im_end|>\n{% endif %}{% endfor %}{% if add_generation_prompt %}<|im_start|>assistant\n{% endif %}"
3
  }
adapter_checkpoint/merges.txt ADDED
The diff for this file is too large to render. See raw diff
 
adapter_checkpoint/preprocessor_config.json CHANGED
@@ -1,29 +1,29 @@
1
  {
2
- "do_convert_rgb": null,
3
  "do_normalize": true,
4
- "do_pan_and_scan": null,
5
  "do_rescale": true,
6
  "do_resize": true,
7
  "image_mean": [
8
- 0.5,
9
- 0.5,
10
- 0.5
11
  ],
12
- "image_processor_type": "Gemma3ImageProcessor",
13
- "image_seq_length": 256,
14
  "image_std": [
15
- 0.5,
16
- 0.5,
17
- 0.5
18
  ],
19
- "pan_and_scan_max_num_crops": null,
20
- "pan_and_scan_min_crop_size": null,
21
- "pan_and_scan_min_ratio_to_activate": null,
22
- "processor_class": "Gemma3Processor",
23
- "resample": 2,
 
24
  "rescale_factor": 0.00392156862745098,
25
  "size": {
26
- "height": 896,
27
- "width": 896
28
- }
 
29
  }
 
1
  {
2
+ "do_convert_rgb": true,
3
  "do_normalize": true,
 
4
  "do_rescale": true,
5
  "do_resize": true,
6
  "image_mean": [
7
+ 0.48145466,
8
+ 0.4578275,
9
+ 0.40821073
10
  ],
11
+ "image_processor_type": "Qwen2VLImageProcessor",
 
12
  "image_std": [
13
+ 0.26862954,
14
+ 0.26130258,
15
+ 0.27577711
16
  ],
17
+ "max_pixels": 12845056,
18
+ "merge_size": 2,
19
+ "min_pixels": 3136,
20
+ "patch_size": 14,
21
+ "processor_class": "Qwen2_5_VLProcessor",
22
+ "resample": 3,
23
  "rescale_factor": 0.00392156862745098,
24
  "size": {
25
+ "longest_edge": 12845056,
26
+ "shortest_edge": 3136
27
+ },
28
+ "temporal_patch_size": 2
29
  }
adapter_checkpoint/special_tokens_map.json CHANGED
@@ -1,30 +1,28 @@
1
  {
2
- "boi_token": "<start_of_image>",
3
- "bos_token": {
4
- "content": "<bos>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- "eoi_token": "<end_of_image>",
 
 
 
 
 
 
11
  "eos_token": {
12
- "content": "<eos>",
13
  "lstrip": false,
14
  "normalized": false,
15
  "rstrip": false,
16
  "single_word": false
17
  },
18
- "image_token": "<image_soft_token>",
19
  "pad_token": {
20
- "content": "<pad>",
21
- "lstrip": false,
22
- "normalized": false,
23
- "rstrip": false,
24
- "single_word": false
25
- },
26
- "unk_token": {
27
- "content": "<unk>",
28
  "lstrip": false,
29
  "normalized": false,
30
  "rstrip": false,
 
1
  {
2
+ "additional_special_tokens": [
3
+ "<|im_start|>",
4
+ "<|im_end|>",
5
+ "<|object_ref_start|>",
6
+ "<|object_ref_end|>",
7
+ "<|box_start|>",
8
+ "<|box_end|>",
9
+ "<|quad_start|>",
10
+ "<|quad_end|>",
11
+ "<|vision_start|>",
12
+ "<|vision_end|>",
13
+ "<|vision_pad|>",
14
+ "<|image_pad|>",
15
+ "<|video_pad|>"
16
+ ],
17
  "eos_token": {
18
+ "content": "<|im_end|>",
19
  "lstrip": false,
20
  "normalized": false,
21
  "rstrip": false,
22
  "single_word": false
23
  },
 
24
  "pad_token": {
25
+ "content": "<|endoftext|>",
 
 
 
 
 
 
 
26
  "lstrip": false,
27
  "normalized": false,
28
  "rstrip": false,
adapter_checkpoint/tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d786405177734910d7a3db625c2826640964a0b4e5cdbbd70620ae3313a01bef
3
- size 33384722
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba0c439f7be467bf47d12a7e6f9adc6116201056fc60c67f431c679b7c16afc8
3
+ size 11422064
adapter_checkpoint/tokenizer_config.json CHANGED
The diff for this file is too large to render. See raw diff
 
adapter_checkpoint/training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d96e298d5f8f669ce7e8ff671c6223b2695ee4aeedd50855b18532acd243a113
3
  size 6033
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3b77494b0fd98379a9da9cec7dcda2471e45cd337a7b27a857c4b62b0c126e8
3
  size 6033
adapter_checkpoint/vocab.json ADDED
The diff for this file is too large to render. See raw diff