--- license: apache-2.0 base_model: - Qwen/Qwen2.5-7B pipeline_tag: text-generation language: - en library_name: transformers tags: - text-generation-inference --- ## Model Description Optimized Layer Merging (OLM) Is a transformer optimization framework implementing automated layer recombination. Olm create Frankenstein's monster out of language models by cherry-picking the best performing layers across different models to create a superior hybrid. The core mechanism: - Takes multiple language models as input - Uses a base model as the foundation - Iteratively replaces individual layers, evaluating performance on specified datasets - Keeps the best performing layer at each position based on metrics like perplexity, exact match, and a custom "quality" score - Builds a fusion model layer-by-layer while maintaining or improving performance https://github.com/jeffmeloy/olm