3rd-Degree-Burn's picture
Update README.md
2bec01c verified
|
raw
history blame
1.31 kB
metadata
license: apache-2.0
tags:
  - merge
  - mergekit
  - lazymergekit
  - NousResearch/Meta-Llama-3.1-8B-Instruct
  - EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta
  - nvidia/OpenMath2-Llama3.1-8B
base_model:
  - NousResearch/Meta-Llama-3.1-8B-Instruct
  - EpistemeAI/Fireball-Alpaca-Llama3.1.07-8B-Philos-Math-KTO-beta
  - nvidia/OpenMath2-Llama3.1-8B

Llama-3.1-8B-Squareroot

This is a TIES merge that combines the performance of the following models:

image/png

Description

I observed that when a model is trained to do just math, it does badly on everything else. So my plan was to merge a “math” model with a strong reasoning/inference model and a general instruction-following model. The result should be a model that's steerable (able to follow instructions) and still good at math.

Examples

Benchmarks

Coming very soon!