nielsr HF Staff commited on
Commit
ffb8dcb
·
verified ·
1 Parent(s): 90b19d0

Improve model card: Update pipeline tag, add dataset, and HF paper link

Browse files

This PR aims to improve the model card by:

- Updating the `pipeline_tag` from `visual-question-answering` to `image-text-to-text` to better reflect the model's capabilities as a Multimodal Large Language Model.
- Adding `MLLM-CL/MLLM-CL-ReplayData` to the `datasets` metadata, as referenced in the project's GitHub README.
- Including the Hugging Face paper link alongside the existing arXiv link for improved accessibility to the paper.

These changes enhance the model's discoverability and provide more comprehensive information for users on the Hugging Face Hub.

Files changed (1) hide show
  1. README.md +11 -10
README.md CHANGED
@@ -1,13 +1,17 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
3
  language:
4
  - en
 
 
5
  metrics:
6
  - accuracy
7
- base_model:
8
- - llava-hf/llava-1.5-7b-hf
9
- - OpenGVLab/InternVL-Chat-ViT-6B-Vicuna-7B
10
- base_model_relation: adapter
11
  tags:
12
  - finance
13
  - medical
@@ -25,10 +29,7 @@ tags:
25
  - multimodal
26
  - image-to-text
27
  - text-generation
28
- pipeline_tag: visual-question-answering
29
- library_name: transformers
30
- datasets:
31
- - MLLM-CL/MLLM-CL
32
  ---
33
 
34
  ## MLLM-CL Benchmark Description
@@ -36,7 +37,7 @@ MLLM-CL is a novel benchmark encompassing domain and ability continual learning,
36
  whereas the latter evaluates on non-IID scenarios with emerging model ability.
37
  For more details, please refer to:
38
 
39
- **MLLM-CL: Continual Learning for Multimodal Large Language Models** [[paper](https://arxiv.org/abs/2506.05453)], [[code](https://github.com/bjzhb666/MLLM-CL/)].
40
  ![](MLLM-CL.png "Magic Gardens")
41
  [‪Hongbo Zhao](https://scholar.google.com/citations?user=Gs22F0UAAAAJ&hl=zh-CN), [Fei Zhu](https://impression2805.github.io/), [Haiyang Guo](https://ghy0501.github.io/guohaiyang0501.github.io/), [Meng Wang](https://moenupa.github.io/), Rundong Wang, [‪Gaofeng Meng](https://scholar.google.com/citations?hl=zh-CN&user=5hti_r0AAAAJ), [‪Zhaoxiang Zhang‬](https://scholar.google.com/citations?hl=zh-CN&user=qxWfV6cAAAAJ)
42
 
 
1
  ---
2
+ base_model:
3
+ - llava-hf/llava-1.5-7b-hf
4
+ - OpenGVLab/InternVL-Chat-ViT-6B-Vicuna-7B
5
+ datasets:
6
+ - MLLM-CL/MLLM-CL
7
+ - MLLM-CL/MLLM-CL-ReplayData
8
  language:
9
  - en
10
+ library_name: transformers
11
+ license: apache-2.0
12
  metrics:
13
  - accuracy
14
+ pipeline_tag: image-text-to-text
 
 
 
15
  tags:
16
  - finance
17
  - medical
 
29
  - multimodal
30
  - image-to-text
31
  - text-generation
32
+ base_model_relation: adapter
 
 
 
33
  ---
34
 
35
  ## MLLM-CL Benchmark Description
 
37
  whereas the latter evaluates on non-IID scenarios with emerging model ability.
38
  For more details, please refer to:
39
 
40
+ **MLLM-CL: Continual Learning for Multimodal Large Language Models** [[paper](https://arxiv.org/abs/2506.05453)], [[HF paper](https://huggingface.co/papers/2506.05453)], [[code](https://github.com/bjzhb666/MLLM-CL/)].
41
  ![](MLLM-CL.png "Magic Gardens")
42
  [‪Hongbo Zhao](https://scholar.google.com/citations?user=Gs22F0UAAAAJ&hl=zh-CN), [Fei Zhu](https://impression2805.github.io/), [Haiyang Guo](https://ghy0501.github.io/guohaiyang0501.github.io/), [Meng Wang](https://moenupa.github.io/), Rundong Wang, [‪Gaofeng Meng](https://scholar.google.com/citations?hl=zh-CN&user=5hti_r0AAAAJ), [‪Zhaoxiang Zhang‬](https://scholar.google.com/citations?hl=zh-CN&user=qxWfV6cAAAAJ)
43