HunminVL 32B v0.1

HunminVL 32B v0.1์€ Qwen2.5-VL-32B-Instruct๋ฅผ ๊ธฐ๋ฐ˜์œผ๋กœ ํ•œ ํ•œ๊ตญ์–ด ์ „์šฉ Vision-Language Instruction ๋ชจ๋ธ์ž…๋‹ˆ๋‹ค.
VL ๋ฐ์ดํ„ฐ๋กœ ๋ณ„๋„ ํ•™์Šต ์—†์ด, **lm_head ๋ ˆ๋ฒจ์—์„œ ์ค‘๊ตญ์–ด, ์ผ๋ณธ์–ด, ๋Ÿฌ์‹œ์•„์–ด์— ํ† ํฐ์„ ๋งˆ์Šคํ‚นํ•˜์—ฌ Language Confusion์„ ์–ต์ œํ•˜๊ณ  ํ•œ๊ตญ์–ด ์ƒ์„ฑ ๋Šฅ๋ ฅ์„ ๋ณด์กดํ•˜๋Š” ๊ฒƒ์„ ๋ชฉํ‘œ๋กœ ํ•ฉ๋‹ˆ๋‹ค.

๐Ÿ”ง ์ฃผ์š” ๋ณ€๊ฒฝ ์‚ฌํ•ญ

  • ์ถ”๊ฐ€์ ์ธ Finetuning ์—†์Œ: ์ถ”๊ฐ€์ ์ธ VL ๋ฐ์ดํ„ฐ ํ•™์Šต ์—†์ด ๊ธฐ์กด Qwen2.5-VL-32B-Instruct ์‚ฌ์šฉ
  • Language Confusion ์™„ํ™”:
    • lm_head์—์„œ ์ค‘๊ตญ์–ด, ์ผ๋ณธ์–ด, ๋Ÿฌ์‹œ์•„์–ด ํ† ํฐ weight์„ masking
    • ํ•™์Šต ์—†์ด ์ถ”๋ก  ์‹œ ์ž๋™ ์ ์šฉ

๐Ÿ“Š Evaluation: Korean VLM Benchmarks

ํ•œ๊ตญ์–ด ์ „์šฉ VLM ํ‰๊ฐ€ ๋ฒค์น˜๋งˆํฌ๋ฅผ ์‚ฌ์šฉํ•ด Language Confusion์„ ์ธก์ •ํ•˜์˜€์Šต๋‹ˆ๋‹ค.
๊ฒฐ๊ณผ๋Š” ๋‹ค์Œ๊ณผ ๊ฐ™์ด ๊ธฐ์กด์˜ ์„ฑ๋Šฅ์„ ์œ ์ง€ํ•˜๋ฉด์„œ๋„ Confusion์ด ์œ ์˜๋ฏธํ•˜๊ฒŒ ์ค„์–ด๋“ฆ์„ ๋ณด์—ฌ์ค๋‹ˆ๋‹ค:

Model Variant K-MMBench K-SEED K-MMStar K-DTCBench K-LLAVA-W
baseline 80.39%
0.12% (5 / 4329)
77.08%
0%
52.60%
0.60% (9 / 1500)
76.67%
0%
89.08%
15.00% (9 / 60)
weight_masking 80.39%
0.02% (1 / 4329)
77.08%
0%
52.53%
0.00% (0 / 1500)
76.67%
0%
87.70%
1.67% (1 / 60)
  • ์ƒ๋‹จ: Performance
  • ํ•˜๋‹จ: Language Confusion ๋น„์œจ (์ค‘๊ตญ์–ด ํ† ํฐ์ด ๋“ฑ์žฅํ•œ ๋น„์œจ)

๐Ÿ’ก Intended Use

  • ํ•œ๊ตญ์–ด ๊ธฐ๋ฐ˜ ๋ฉ€ํ‹ฐ๋ชจ๋‹ฌ ์–ด์‹œ์Šคํ„ดํŠธ
  • ํ•œ๊ตญ์–ด VQA ๋ฐ ์ด๋ฏธ์ง€ ์„ค๋ช…
  • ํ•œ๊ตญ์–ด instruction-following with image

โš ๏ธ Limitations

  • ํ•™์Šต ๋ฐ์ดํ„ฐ ์—†์ด inference-only masking์ด๋ฏ€๋กœ ๊ทนํ•œ ์ƒํ™ฉ์—์„œ๋Š” ์ œํ•œ์ 
  • ์ค‘๊ตญ์–ด ๊ฐ•์ œ ์ƒ์„ฑ์€ ๋น„์ •์ƒ์ ์œผ๋กœ ์–ต์ œ๋  ์ˆ˜ ์žˆ์Œ

๐Ÿ’ป ์˜ˆ์‹œ ์‚ฌ์šฉ๋ฒ•

from transformers import Qwen2_5_VLForConditionalGeneration, AutoProcessor

model = Qwen2_5_VLForConditionalGeneration.from_pretrained("mncai/hunminVL_32B_v0.1_20250724", device_map="auto")
processor = AutoProcessor.from_pretrained("mncai/hunminVL_32B_v0.1_20250724")

# ์ž…๋ ฅ ์˜ˆ์‹œ: vision + text
Downloads last month
15
Safetensors
Model size
33.5B params
Tensor type
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support