-
One-RL-to-See-Them-All/Orsta-7B
Image-Text-to-Text • Updated • 626 • 7 -
One-RL-to-See-Them-All/Orsta-32B-0321
Image-Text-to-Text • Updated • 18 -
One-RL-to-See-Them-All/Orsta-32B-0326
Image-Text-to-Text • Updated • 125 • 4 -
One RL to See Them All: Visual Triple Unified Reinforcement Learning
Paper • 2505.18129 • Published • 59
Collections
Discover the best community collections!
Collections including paper arxiv:2505.18129
-
EVA-CLIP-18B: Scaling CLIP to 18 Billion Parameters
Paper • 2402.04252 • Published • 29 -
Vision Superalignment: Weak-to-Strong Generalization for Vision Foundation Models
Paper • 2402.03749 • Published • 13 -
ScreenAI: A Vision-Language Model for UI and Infographics Understanding
Paper • 2402.04615 • Published • 44 -
EfficientViT-SAM: Accelerated Segment Anything Model Without Performance Loss
Paper • 2402.05008 • Published • 23
-
RL + Transformer = A General-Purpose Problem Solver
Paper • 2501.14176 • Published • 28 -
Towards General-Purpose Model-Free Reinforcement Learning
Paper • 2501.16142 • Published • 30 -
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training
Paper • 2501.17161 • Published • 122 -
MaxInfoRL: Boosting exploration in reinforcement learning through information gain maximization
Paper • 2412.12098 • Published • 5
-
MLLM-as-a-Judge for Image Safety without Human Labeling
Paper • 2501.00192 • Published • 31 -
2.5 Years in Class: A Multimodal Textbook for Vision-Language Pretraining
Paper • 2501.00958 • Published • 107 -
Xmodel-2 Technical Report
Paper • 2412.19638 • Published • 27 -
HuatuoGPT-o1, Towards Medical Complex Reasoning with LLMs
Paper • 2412.18925 • Published • 105
-
One-RL-to-See-Them-All/Orsta-7B
Image-Text-to-Text • Updated • 626 • 7 -
One-RL-to-See-Them-All/Orsta-32B-0321
Image-Text-to-Text • Updated • 18 -
One-RL-to-See-Them-All/Orsta-32B-0326
Image-Text-to-Text • Updated • 125 • 4 -
One RL to See Them All: Visual Triple Unified Reinforcement Learning
Paper • 2505.18129 • Published • 59
-
One-RL-to-See-Them-All/Orsta-7B
Image-Text-to-Text • Updated • 626 • 7 -
One-RL-to-See-Them-All/Orsta-32B-0321
Image-Text-to-Text • Updated • 18 -
One-RL-to-See-Them-All/Orsta-32B-0326
Image-Text-to-Text • Updated • 125 • 4 -
One RL to See Them All: Visual Triple Unified Reinforcement Learning
Paper • 2505.18129 • Published • 59
-
BLIP3-o: A Family of Fully Open Unified Multimodal Models-Architecture, Training and Dataset
Paper • 2505.09568 • Published • 90 -
Qwen3 Technical Report
Paper • 2505.09388 • Published • 182 -
GuardReasoner-VL: Safeguarding VLMs via Reinforced Reasoning
Paper • 2505.11049 • Published • 59 -
Emerging Properties in Unified Multimodal Pretraining
Paper • 2505.14683 • Published • 129
-
Unified Multimodal Understanding and Generation Models: Advances, Challenges, and Opportunities
Paper • 2505.02567 • Published • 74 -
TabSTAR: A Foundation Tabular Model With Semantically Target-Aware Representations
Paper • 2505.18125 • Published • 109 -
Distilling LLM Agent into Small Models with Retrieval and Code Tools
Paper • 2505.17612 • Published • 76 -
One RL to See Them All: Visual Triple Unified Reinforcement Learning
Paper • 2505.18129 • Published • 59
-
CLIMB: CLustering-based Iterative Data Mixture Bootstrapping for Language Model Pre-training
Paper • 2504.13161 • Published • 92 -
Hebbian Learning based Orthogonal Projection for Continual Learning of Spiking Neural Networks
Paper • 2402.11984 • Published -
BlackGoose Rimer: Harnessing RWKV-7 as a Simple yet Superior Replacement for Transformers in Large-Scale Time Series Modeling
Paper • 2503.06121 • Published • 5 -
Timer: Transformers for Time Series Analysis at Scale
Paper • 2402.02368 • Published