Update README.md
Browse files
README.md
CHANGED
|
@@ -10,26 +10,29 @@ license: cc-by-nc-4.0
|
|
| 10 |
|
| 11 |
OpenOrca-Platypus2-13B is a merge of [`garage-bAInd/Platypus2-13B`](https://huggingface.co/garage-bAInd/Platypus2-13B) and [`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B).
|
| 12 |
|
|
|
|
|
|
|
| 13 |

|
| 14 |
|
| 15 |
### Benchmark Metrics
|
| 16 |
|
| 17 |
| Metric | Value |
|
| 18 |
|-----------------------|-------|
|
| 19 |
-
| MMLU (5-shot) |
|
| 20 |
-
| ARC (25-shot) |
|
| 21 |
-
| HellaSwag (10-shot) |
|
| 22 |
-
| TruthfulQA (0-shot) |
|
| 23 |
-
| Avg. |
|
| 24 |
|
| 25 |
We use state-of-the-art [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) to run the benchmark tests above, using the same version as the HuggingFace LLM Leaderboard. Please see below for detailed instructions on reproducing benchmark results.
|
| 26 |
|
| 27 |
### Model Details
|
| 28 |
|
| 29 |
-
* **Trained by**: **Platypus2-13B** trained by Cole Hunter & Ariel Lee; **OpenOrcaxOpenChat-Preview2-13B** trained by
|
| 30 |
* **Model type:** **OpenOrca-Platypus2-13B** is an auto-regressive language model based on the LLaMA 2 transformer architecture.
|
| 31 |
* **Language(s)**: English
|
| 32 |
* **License for Platypus2-13B base weights**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
|
|
|
|
| 33 |
|
| 34 |
### Prompt Template for base Platypus2-13B
|
| 35 |
```
|
|
@@ -39,13 +42,17 @@ We use state-of-the-art [Language Model Evaluation Harness](https://github.com/E
|
|
| 39 |
|
| 40 |
### Response:
|
| 41 |
```
|
|
|
|
|
|
|
| 42 |
|
| 43 |
-
### Training
|
| 44 |
|
| 45 |
`garage-bAInd/Platypus2-13B` trained using STEM and logic based dataset [`garage-bAInd/Open-Platypus`](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
|
| 46 |
|
| 47 |
Please see our [paper](https://platypus-llm.github.io/Platypus.pdf) and [project webpage](https://platypus-llm.github.io) for additional information.
|
| 48 |
|
|
|
|
|
|
|
| 49 |
### Training Procedure
|
| 50 |
|
| 51 |
`garage-bAInd/Platypus2-13B` was instruction fine-tuned using LoRA on 1 A100 80GB. For training details and inference instructions please see the [Platypus](https://github.com/arielnlee/Platypus) GitHub repo.
|
|
|
|
| 10 |
|
| 11 |
OpenOrca-Platypus2-13B is a merge of [`garage-bAInd/Platypus2-13B`](https://huggingface.co/garage-bAInd/Platypus2-13B) and [`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B).
|
| 12 |
|
| 13 |
+
Thank you Open-Orca for putting out a beast of a model and dataset—can't wait for the 70B version (and beyond)!
|
| 14 |
+
|
| 15 |

|
| 16 |
|
| 17 |
### Benchmark Metrics
|
| 18 |
|
| 19 |
| Metric | Value |
|
| 20 |
|-----------------------|-------|
|
| 21 |
+
| MMLU (5-shot) | 59.5 |
|
| 22 |
+
| ARC (25-shot) | 62.88 |
|
| 23 |
+
| HellaSwag (10-shot) | 83.19 |
|
| 24 |
+
| TruthfulQA (0-shot) | 52.69 |
|
| 25 |
+
| Avg. | 64.56 |
|
| 26 |
|
| 27 |
We use state-of-the-art [Language Model Evaluation Harness](https://github.com/EleutherAI/lm-evaluation-harness) to run the benchmark tests above, using the same version as the HuggingFace LLM Leaderboard. Please see below for detailed instructions on reproducing benchmark results.
|
| 28 |
|
| 29 |
### Model Details
|
| 30 |
|
| 31 |
+
* **Trained by**: **Platypus2-13B** trained by Cole Hunter & Ariel Lee; **OpenOrcaxOpenChat-Preview2-13B** trained by Open-Orca
|
| 32 |
* **Model type:** **OpenOrca-Platypus2-13B** is an auto-regressive language model based on the LLaMA 2 transformer architecture.
|
| 33 |
* **Language(s)**: English
|
| 34 |
* **License for Platypus2-13B base weights**: Non-Commercial Creative Commons license ([CC BY-NC-4.0](https://creativecommons.org/licenses/by-nc/4.0/))
|
| 35 |
+
* **License for OpenOrcaxOpenChat-Preview2-13B base weights**: LLaMa-2 commercial
|
| 36 |
|
| 37 |
### Prompt Template for base Platypus2-13B
|
| 38 |
```
|
|
|
|
| 42 |
|
| 43 |
### Response:
|
| 44 |
```
|
| 45 |
+
### Prompt Template for base OpenOrcaxOpenChat-Preview2-13B
|
| 46 |
+
OpenChat Llama2 V1: see [Open-Orca's page](https://huggingface.co/Open-Orca/OpenOrcaxOpenChat-Preview2-13B) for additional information.
|
| 47 |
|
| 48 |
+
### Training Datasets
|
| 49 |
|
| 50 |
`garage-bAInd/Platypus2-13B` trained using STEM and logic based dataset [`garage-bAInd/Open-Platypus`](https://huggingface.co/datasets/garage-bAInd/Open-Platypus).
|
| 51 |
|
| 52 |
Please see our [paper](https://platypus-llm.github.io/Platypus.pdf) and [project webpage](https://platypus-llm.github.io) for additional information.
|
| 53 |
|
| 54 |
+
[`Open-Orca/OpenOrcaxOpenChat-Preview2-13B`] trained using a refined, 220k subset of the [OpenOrca dataset](https://huggingface.co/datasets/Open-Orca/OpenOrca).
|
| 55 |
+
|
| 56 |
### Training Procedure
|
| 57 |
|
| 58 |
`garage-bAInd/Platypus2-13B` was instruction fine-tuned using LoRA on 1 A100 80GB. For training details and inference instructions please see the [Platypus](https://github.com/arielnlee/Platypus) GitHub repo.
|