ModernVBERT
AI & ML interests
None defined yet.
Recent Activity
ModernVBERT: Towards Smaller Visual Document Retrievers

This organization contains all artifacts released with our preprint ModernVBERT: Towards Smaller Visual Document Retrievers.
Abstract
Multimodal embedding models are gaining prevalence, notably for document retrieval as efficient alternatives to text-only pipelines. These models are typically built by finetuning large vision–language decoders (VLMs) with contrastive losses on text–image pairs. In this work, we show that, while cost-efficient, this repurposing approach often bottlenecks retrieval performance. Through controlled experiments, we establish a principled recipe for improving visual document retrieval models. We notably measure the impact of attention masking, image resolution, modality alignment data regimes, and late interaction centered contrastive objectives which emerge as central performance factors. Building on these insights, we release ModernVBERT, a compact 250M-parameter vision–language encoder that outperforms models up to 10 times larger when finetuned on document retrieval tasks.
Ressources
- HuggingFace Project Page: The HF page centralizing everything!
- Preprint: The paper with all details!
- Blog Post: The blog post introducing our release!
- Finetuning Tutorial: A colab notebook to learn how to finetune ModernVBERT for document retrieval!
- Code: Released soon...
Contact of the authors
- Paul Teiletche: [email protected]
- Quentin Macé: [email protected]
- Max Conti: [email protected]
- Manuel Faysse: [email protected]
Citation
If you use any datasets or models from this organization in your research, please cite the original dataset as follows:
@misc{teiletche2025modernvbertsmallervisualdocument,
title={ModernVBERT: Towards Smaller Visual Document Retrievers},
author={Paul Teiletche and Quentin Macé and Max Conti and Antonio Loison and Gautier Viaud and Pierre Colombo and Manuel Faysse},
year={2025},
eprint={2510.01149},
archivePrefix={arXiv},
primaryClass={cs.IR},
url={https://arxiv.org/abs/2510.01149},
}
Acknowledgments
This work was carried out within the framework of the LIAGORA "LabCom", a joint laboratory supported by the French National Research Agency (ANR) and established between ILLUIN Technology and the MICS laboratory of CentraleSupélec. This work was performed using HPC resources from IDRIS with grant AD011016393. We warmly thank Hippolyte Gisserot-Boukhlef and Nicolas Boizard for sharing the controlled experiments LM checkpoints, Antoine Chaffin for his feedback on the modality alignment codebase and insights on Ettin’s modeling, as well as Andi Marafioti, Orr Zohar, and Miquel Farré for their valuable input and help on gathering the modality alignment dataset.
models
5

ModernVBERT/modernvbert

ModernVBERT/bimodernvbert

ModernVBERT/modernvbert-embed

ModernVBERT/colmodernvbert-base
