Delta Activations: A Representation for Finetuned Large Language Models
Abstract
Delta Activations represent fine-tuned models as vector embeddings based on internal activation shifts, enabling effective clustering and model reuse.
The success of powerful open source Large Language Models (LLMs) has enabled the community to create a vast collection of post-trained models adapted to specific tasks and domains. However, navigating and understanding these models remains challenging due to inconsistent metadata and unstructured repositories. We introduce Delta Activations, a method to represent finetuned models as vector embeddings by measuring shifts in their internal activations relative to a base model. This representation allows for effective clustering by domain and task, revealing structure in the model landscape. Delta Activations also demonstrate desirable properties: it is robust across finetuning settings and exhibits an additive property when finetuning datasets are mixed. In addition, we show that Delta Activations can embed tasks via few-shot finetuning, and further explore its use for model selection and merging. We hope Delta Activations can facilitate the practice of reusing publicly available models. Code is available at https://github.com/OscarXZQ/delta_activations.
Community
⚙️Code: https://github.com/OscarXZQ/delta_activations
🗺️Project page and navigator: https://oscarxzq.github.io/delta_activation/
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- Causal2Vec: Improving Decoder-only LLMs as Versatile Embedding Models (2025)
- Resource-Efficient Adaptation of Large Language Models for Text Embeddings via Prompt Engineering and Contrastive Fine-tuning (2025)
- MSGCoOp: Multiple Semantic-Guided Context Optimization for Few-Shot Learning (2025)
- Soft Injection of Task Embeddings Outperforms Prompt-Based In-Context Learning (2025)
- LMAR: Language Model Augmented Retriever for Domain-specific Knowledge Indexing (2025)
- EmbedGrad: Gradient-Based Prompt Optimization in Embedding Space for Large Language Models (2025)
- Enhancing Cross-task Transfer of Large Language Models via Activation Steering (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper