
OpenGVLab
community
AI & ML interests
Computer Vision
Recent Activity
Organization Card
OpenGVLab
Welcome to OpenGVLab! We are a research group from Shanghai AI Lab focused on Vision-Centric AI research. The GV in our name, OpenGVLab, means general vision, a general understanding of vision, so little effort is needed to adapt to new vision-based tasks.
Models
- InternVL: a pioneering open-source alternative to GPT-4V.
- InternImage: a large-scale vision foundation models with deformable convolutions.
- InternVideo: large-scale video foundation models for multimodal understanding.
- VideoChat: an end-to-end chat assistant for video comprehension.
- All-Seeing-Project: towards panoptic visual recognition and understanding of the open world.
Datasets
- ShareGPT4o: a groundbreaking large-scale resource that we plan to open-source with 200K meticulously annotated images, 10K videos with highly descriptive captions, and 10K audio files with detailed descriptions.
- InternVid: a large-scale video-text dataset for multimodal understanding and generation.
- MMPR: a high-quality, large-scale multimodal preference dataset.
Benchmarks
- MVBench: a comprehensive benchmark for multimodal video understanding.
- CRPE: a benchmark covering all elements of the relation triplets (subject, predicate, object), providing a systematic platform for the evaluation of relation comprehension ability.
- MM-NIAH: a comprehensive benchmark for long multimodal documents comprehension.
- GMAI-MMBench: a comprehensive multimodal evaluation benchmark towards general medical AI.
Collections
22
spaces
11
Runtime error
InternVideo2.5
💬
Hierarchical Compression for Long-Context Video Modeling
Running
446
InternVL
⚡
Chat with an AI that understands text and images
Running
37
MVBench Leaderboard
🐨
Submit model evaluation and view leaderboard
Running
on
Zero
16
InternVideo2 Chat 8B HD
👁
Upload a video to chat about its contents
Running
10
ControlLLM
🚀
Display maintenance message for ControlLLM
Running
on
Zero
94
VideoMamba
🐍
Classify video and image content
models
173

OpenGVLab/InternVL3-2B
Image-Text-to-Text
•
Updated
•
654
•
9

OpenGVLab/InternVL3-8B
Image-Text-to-Text
•
Updated
•
439
•
13

OpenGVLab/InternVL3-14B
Image-Text-to-Text
•
Updated
•
654
•
14

OpenGVLab/InternVL3-38B
Image-Text-to-Text
•
Updated
•
364
•
13

OpenGVLab/InternVL3-9B
Image-Text-to-Text
•
Updated
•
102
•
12

OpenGVLab/InternVL3-1B
Image-Text-to-Text
•
Updated
•
468
•
17

OpenGVLab/InternVL3-78B
Image-Text-to-Text
•
Updated
•
2.02k
•
55

OpenGVLab/VisualPRM-8B
Image-Text-to-Text
•
Updated
•
446
•
11

OpenGVLab/Mini-InternVL2-2B-DA-BDD
Image-Text-to-Text
•
Updated
•
29
•
1

OpenGVLab/Mini-InternVL2-2B-DA-DriveLM
Image-Text-to-Text
•
Updated
•
30
datasets
38
OpenGVLab/MMPR-v1.2-prompts
Updated
•
12
•
1
OpenGVLab/MMPR-v1.1
Preview
•
Updated
•
560
•
46
OpenGVLab/MMPR-v1.2
Updated
•
18
•
4
OpenGVLab/MMPR
Preview
•
Updated
•
98
•
49
OpenGVLab/LongVid
Preview
•
Updated
•
11
•
1
OpenGVLab/NIAH-Video
Viewer
•
Updated
•
629
•
85
OpenGVLab/OmniCorpus-CC-210M
Viewer
•
Updated
•
208M
•
622
•
23
OpenGVLab/OmniCorpus-YT
Updated
•
445
•
12
OpenGVLab/OmniCorpus-CC
Viewer
•
Updated
•
872M
•
14.5k
•
16
OpenGVLab/VisualPRM400K
Preview
•
Updated
•
386
•
7