Model Description

Optimized Layer Merging (OLM) Is a transformer optimization framework implementing automated layer recombination.

Olm create Frankenstein's monster out of language models by cherry-picking the best performing layers across different models to create a superior hybrid. The core mechanism:

  • Takes multiple language models as input
  • Uses a base model as the foundation
  • Iteratively replaces individual layers, evaluating performance on specified datasets
  • Keeps the best performing layer at each position based on metrics like perplexity, exact match, and a custom "quality" score
  • Builds a fusion model layer-by-layer while maintaining or improving performance

https://github.com/jeffmeloy/olm

Downloads last month
12
Safetensors
Model size
7.62B params
Tensor type
BF16
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Model tree for jeffmeloy/Qwen2.5-7B-olm-v1.2

Base model

Qwen/Qwen2.5-7B
Finetuned
(177)
this model
Quantizations
3 models