Model Description

Optimized Layer Merging (OLM) Is a transformer optimization framework implementing automated layer recombination.

Olm create Frankenstein's monster out of language models by cherry-picking the best performing layers across different models to create a superior hybrid. The core mechanism:

  • Takes multiple language models as input
  • Uses a base model as the foundation
  • Iteratively replaces individual layers, evaluating performance on specified datasets
  • Keeps the best performing layer at each position based on metrics like perplexity, exact match, and a custom "quality" score
  • Builds a fusion model layer-by-layer while maintaining or improving performance

https://github.com/jeffmeloy/olm

Downloads last month
10
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for jeffmeloy/Qwen2.5-7B-olm-v1.4

Base model

Qwen/Qwen2.5-7B
Finetuned
(196)
this model
Quantizations
2 models