Update pipeline tag and add library name (#1)
Browse files- Update pipeline tag and add library name (f87475b5c42b2880f9c6e72a38cbf78c467c3d2e)
Co-authored-by: Niels Rogge <[email protected]>
README.md
CHANGED
@@ -1,17 +1,17 @@
|
|
1 |
---
|
2 |
-
|
|
|
3 |
datasets:
|
4 |
- antgroup/HumanSense_Benchmark
|
5 |
language:
|
6 |
- en
|
|
|
7 |
metrics:
|
8 |
- accuracy
|
9 |
-
|
10 |
-
|
11 |
-
pipeline_tag: visual-question-answering
|
12 |
---
|
13 |
|
14 |
-
|
15 |
<div align="center" style="font-family: charter;">
|
16 |
|
17 |
|
@@ -22,131 +22,4 @@ pipeline_tag: visual-question-answering
|
|
22 |
<!-- <h1></br>From Multimodal Perception to Empathetic Context-Aware Responses through Reasoning MLLMs</h1> -->
|
23 |
|
24 |
<div>
|
25 |
-
<a href="https://scholar.google.com/citations?user=
|
26 |
-
<a href="https://scholar.google.com/citations?user=S8FmqTUAAAAJ&hl=en">Ruobing Zheng<sup>*</sup><sup>2</sup></a>,
|
27 |
-
<a href="https://scholar.google.com/citations?user=3WVFdMUAAAAJ&hl=en">Yabing Wang<sup>1</sup></a>,
|
28 |
-
<a href="https://scholar.google.com/citations?user=yOtsVWQAAAAJ&hl=en&oi=sra">Tianqi Li<sup>2</sup></a>,
|
29 |
-
<a href="https://yuanyi.pub/">Yi Yuan<sup>2</sup></a>,
|
30 |
-
<a href="https://scholar.google.com/citations?hl=en&user=8SCEv-YAAAAJ&view_op=list_works&sortby=pubdate">Jingdong Chen<sup>2</sup></a>,
|
31 |
-
<a href="https://scholar.google.com/citations?user=RypRCUQAAAAJ&hl=en">Le Wang<sup>†<dag><sup>1</sup></a> <br>
|
32 |
-
<span style="font-size: 13px; margin-top: 0.8em">
|
33 |
-
<br>
|
34 |
-
<sup>*</sup>Co-first authors. Project Lead.
|
35 |
-
<sup>†</sup>Corresponding Author.
|
36 |
-
<br>
|
37 |
-
<sup>1</sup>Xi’an Jiaotong University. <sup>2</sup>Ant Group.
|
38 |
-
<br>
|
39 |
-
</span>
|
40 |
-
</div>
|
41 |
-
|
42 |
-
|
43 |
-
<a target="_blank" href="https://arxiv.org/abs/2508.10576" ><button><i class="ai ai-arxiv"></i> arXiv:2508.10576</button></a>
|
44 |
-
|
45 |
-
<a target="_blank" href="https://digital-avatar.github.io/ai/HumanSense/" ><button><i class="ai ai-arxiv"></i> Homepage</button></a>
|
46 |
-
|
47 |
-
<a target="_blank" href="https://github.com/antgroup/HumanSense" ><button><i class="ai ai-arxiv"></i> GitHub</button></a>
|
48 |
-
|
49 |
-
|
50 |
-
|
51 |
-
|
52 |
-
|
53 |
-
<img src="figure1.png" width="100%"/>
|
54 |
-
<p align="justify"><i>While Multimodal Large Language Models (MLLMs) show immense promise for achieving truly human-like interactions, progress is hindered by the lack of fine-grained evaluation frameworks for human-centered scenarios, encompassing both the understanding of complex human intentions and the provision of empathetic, context-aware responses. Here we introduce <strong>HumanSense</strong>, a comprehensive benchmark designed to evaluate the human-centered perception and interaction capabilities of MLLMs, with a particular focus on deep understanding of extended multimodal contexts and the formulation of rational feedback. Our evaluation reveals that leading MLLMs still have considerable room for improvement, particularly for advanced interaction-oriented tasks. Supplementing visual input with audio and text information yields substantial improvements, and Omni-modal models show advantages on these tasks. Furthermore, we argue that appropriate feedback stems from a contextual analysis of the interlocutor's needs and emotions, with reasoning ability serving as the key to unlocking it. Accordingly, we devise a multi-stage, modality-progressive reinforcement learning approach, resulting in <strong>HumanSense-Omni-Reasoning</strong>, which substantially enhances performance on higher-level understanding and interactive tasks. Additionally, we observe that successful reasoning processes exhibit highly consistent thought patterns. By designing corresponding prompts, we also enhance the performance of non-reasoning models in a training-free manner.
|
55 |
-
</i></p>
|
56 |
-
|
57 |
-
</div>
|
58 |
-
|
59 |
-
## Release
|
60 |
-
- `2025-08-27` :hearts: We release both the training code and dataset!
|
61 |
-
- `2025-08-27` :hearts: We released Benchmark and code!
|
62 |
-
- `2025-08-15` :rocket: We released our paper!
|
63 |
-
|
64 |
-
|
65 |
-
|
66 |
-
## Quickstart
|
67 |
-
|
68 |
-
Below, we provide simple examples to show how to use HumanSense_Omni_Reasoning with 🤗 Transformers.
|
69 |
-
```
|
70 |
-
pip uninstall transformers
|
71 |
-
pip install transformers==4.52.0
|
72 |
-
pip install accelerate
|
73 |
-
pip install qwen-omni-utils
|
74 |
-
pip install qwen-omni-utils[decord] -U
|
75 |
-
```
|
76 |
-
|
77 |
-
```python
|
78 |
-
import torch
|
79 |
-
from transformers import Qwen2_5OmniForConditionalGeneration, Qwen2_5OmniProcessor
|
80 |
-
from qwen_omni_utils import process_mm_info
|
81 |
-
|
82 |
-
|
83 |
-
model_path = "antgroup/HumanSense_Omni_Reasoning"
|
84 |
-
|
85 |
-
|
86 |
-
model = Qwen2_5OmniForConditionalGeneration.from_pretrained(
|
87 |
-
model_path,
|
88 |
-
torch_dtype=torch.bfloat16,
|
89 |
-
device_map="auto",
|
90 |
-
attn_implementation="flash_attention_2",
|
91 |
-
)
|
92 |
-
model.disable_talker()
|
93 |
-
processor = Qwen2_5OmniProcessor.from_pretrained(model_path)
|
94 |
-
|
95 |
-
conversation = [
|
96 |
-
{
|
97 |
-
"role": "user",
|
98 |
-
"content": [
|
99 |
-
{
|
100 |
-
"type": "video",
|
101 |
-
"video": "file:///path/to/xxx.mp4",
|
102 |
-
"max_pixels": 151200
|
103 |
-
},
|
104 |
-
{
|
105 |
-
"type": "text",
|
106 |
-
"text": "xxxxxxxxxxxxxxxxxx\n"
|
107 |
-
}
|
108 |
-
],
|
109 |
-
}
|
110 |
-
]
|
111 |
-
|
112 |
-
USE_AUDIO_IN_VIDEO=True
|
113 |
-
|
114 |
-
text = processor.apply_chat_template(conversation, add_generation_prompt=True, tokenize=False)
|
115 |
-
audios, images, videos = process_mm_info(conversation, use_audio_in_video=USE_AUDIO_IN_VIDEO)
|
116 |
-
inputs = processor(text=text, audio=audios, images=images, videos=videos, return_tensors="pt", padding=True,padding_side="left",add_special_tokens=False, use_audio_in_video=USE_AUDIO_IN_VIDEO)
|
117 |
-
inputs = inputs.to(model.device).to(model.dtype)
|
118 |
-
|
119 |
-
# Inference: Generation of the output text and audio
|
120 |
-
text_ids = model.generate(**inputs,return_audio=False, use_audio_in_video=USE_AUDIO_IN_VIDEO)
|
121 |
-
generated_ids_trimmed = [
|
122 |
-
out_ids[len(in_ids) :] for in_ids, out_ids in zip(inputs.input_ids, text_ids)
|
123 |
-
]
|
124 |
-
text = processor.batch_decode(generated_ids_trimmed, skip_special_tokens=True, clean_up_tokenization_spaces=False)
|
125 |
-
|
126 |
-
|
127 |
-
|
128 |
-
response = text[0]
|
129 |
-
print('*'*30)
|
130 |
-
print(response)
|
131 |
-
|
132 |
-
```
|
133 |
-
|
134 |
-
<p align="justify"><i>Examples of Reasoning: </i></p>
|
135 |
-
<img src="figure5.png" width="100%"/>
|
136 |
-
<p align="justify"><i>These cases cover four high-level perception and interaction tasks, including both video-based and audio-based questions. The reasoning processes all demonstrate thinking that integrates characteristics, emotions, and context, and then provides appropriate feedback.
|
137 |
-
</i></p>
|
138 |
-
|
139 |
-
</div>
|
140 |
-
|
141 |
-
|
142 |
-
|
143 |
-
**BibTeX:**
|
144 |
-
|
145 |
-
```
|
146 |
-
@article{qin2025humansense,
|
147 |
-
title={HumanSense: From Multimodal Perception to Empathetic Context-Aware Responses through Reasoning MLLMs},
|
148 |
-
author={Qin, Zheng and Zheng, Ruobing and Wang, Yabing and Li, Tianqi and Yuan, Yi and Chen, Jingdong and Wang, Le},
|
149 |
-
journal={arXiv preprint arXiv:2508.10576},
|
150 |
-
year={2025}
|
151 |
-
}
|
152 |
-
```
|
|
|
1 |
---
|
2 |
+
base_model:
|
3 |
+
- Qwen/Qwen2.5-Omni-7B
|
4 |
datasets:
|
5 |
- antgroup/HumanSense_Benchmark
|
6 |
language:
|
7 |
- en
|
8 |
+
license: apache-2.0
|
9 |
metrics:
|
10 |
- accuracy
|
11 |
+
pipeline_tag: video-text-to-text
|
12 |
+
library_name: transformers
|
|
|
13 |
---
|
14 |
|
|
|
15 |
<div align="center" style="font-family: charter;">
|
16 |
|
17 |
|
|
|
22 |
<!-- <h1></br>From Multimodal Perception to Empathetic Context-Aware Responses through Reasoning MLLMs</h1> -->
|
23 |
|
24 |
<div>
|
25 |
+
<a href="https://scholar.google.com/citations?user=sPQ
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|