FuseLIP: Multimodal Embeddings via Early Fusion of Discrete Tokens
Abstract
FuseLIP is a transformer-based architecture that uses a shared vocabulary for text and image tokens to enhance multimodal embedding and outperforms existing models in tasks such as VQA and text-guided image retrieval.
Contrastive language-image pre-training aligns the features of text-image pairs in a common latent space via distinct encoders for each modality. While this approach achieves impressive performance in several zero-shot tasks, it cannot natively handle multimodal inputs, i.e., encoding image and text into a single feature vector. As a remedy, it is common practice to use additional modules to merge the features extracted by the unimodal encoders. In this work, we present FuseLIP, an alternative architecture for multimodal embedding. Leveraging recent progress in discrete image tokenizers, we propose to use a single transformer model which operates on an extended vocabulary of text and image tokens. This early fusion approach allows the different modalities to interact at each depth of encoding and obtain richer representations compared to common late fusion. We collect new datasets for multimodal pre-training and evaluation, designing challenging tasks for multimodal encoder models. We show that FuseLIP outperforms other approaches in multimodal embedding tasks such as VQA and text-guided image transformation retrieval, while being comparable to baselines on unimodal tasks.
Community
FuseLIP: Multimodal Embeddings via Early Fusion of Discrete Tokens
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Breaking the Modality Barrier: Universal Embedding Learning with Multimodal LLMs (2025)
- Learning Joint ID-Textual Representation for ID-Preserving Image Synthesis (2025)
- UniMoCo: Unified Modality Completion for Robust Multi-Modal Embeddings (2025)
- TokLIP: Marry Visual Tokens to CLIP for Multimodal Comprehension and Generation (2025)
- DeepMLF: Multimodal language model with learnable tokens for deep fusion in sentiment analysis (2025)
- Emerging Properties in Unified Multimodal Pretraining (2025)
- CAV-MAE Sync: Improving Contrastive Audio-Visual Mask Autoencoders via Fine-Grained Alignment (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 4
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper