File size: 887 Bytes
f48d6dd
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
---
license: apache-2.0
base_model:
- Qwen/Qwen2.5-7B
pipeline_tag: text-generation
language:
- en
library_name: transformers
tags:
- text-generation-inference
---

## Model Description

Optimized Layer Merging (OLM)
Is a transformer optimization framework implementing automated layer recombination.

Olm create Frankenstein's monster out of language models by cherry-picking the best performing layers across different models to create a superior hybrid.
The core mechanism:

- Takes multiple language models as input
- Uses a base model as the foundation
- Iteratively replaces individual layers, evaluating performance on specified datasets
- Keeps the best performing layer at each position based on metrics like perplexity, exact match, and a custom "quality" score
- Builds a fusion model layer-by-layer while maintaining or improving performance

https://github.com/jeffmeloy/olm