| license: apache-2.0 | |
| tags: | |
| - merge | |
| - mergekit | |
| - lazymergekit | |
| - NousResearch/Meta-Llama-3.1-8B-Instruct | |
| - EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta | |
| - nvidia/OpenMath2-Llama3.1-8B | |
| # Llama-3.1-8B-Squareroot | |
| Llama-3.1-8B-Squareroot is a merge of the following models using [mergekit](https://github.com/cg123/mergekit): | |
| * [NousResearch/Meta-Llama-3.1-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3.1-8B-Instruct) | |
| * [EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta](https://huggingface.co/EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta) | |
| * [nvidia/OpenMath2-Llama3.1-8B](https://huggingface.co/nvidia/OpenMath2-Llama3.1-8B) | |
| ## 🧩 Configuration | |
| ```yaml | |
| models: | |
| - model: NousResearch/Meta-Llama-3.1-8B-Instruct # instruct model | |
| parameters: | |
| density: [1, 0.8, 0.5] | |
| weight: 1.0 | |
| - model: EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta # Reasoning/Logic model | |
| parameters: | |
| density: [0.9, 0.6, 0.4] | |
| weight: [0.3, 0.6, 0.9] | |
| - model: nvidia/OpenMath2-Llama3.1-8B # Math model | |
| parameters: | |
| density: 0.7 | |
| weight: | |
| - filter: mlp | |
| value: 0.7 | |
| - value: 0.3 | |
| merge_method: ties | |
| base_model: vicgalle/Configurable-Llama-3.1-8B-Instruct # General base model | |
| parameters: | |
| normalize: true | |
| dtype: float16 | |
| ``` |