JDBMG commited on
Commit
c8be571
·
verified ·
1 Parent(s): 3754578

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +53 -2
README.md CHANGED
@@ -1,5 +1,56 @@
1
  ---
2
- license: mit
3
- base_model: mistralai/Mistral-7B-v0.1
 
 
 
 
 
 
 
4
 
5
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - mistralai/Mistral-7B-v0.1
4
+ - cognitivecomputations/dolphin-2.2.1-mistral-7b
5
+ - HuggingFaceH4/zephyr-7b-beta
6
+ - NousResearch/Hermes-2-Pro-Mistral-7B
7
+ library_name: transformers
8
+ tags:
9
+ - mergekit
10
+ - merge
11
 
12
  ---
13
+ # Herdolphyr
14
+
15
+ This is a quantitized merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
+
17
+ ## Merge Details
18
+ ### Merge Method
19
+
20
+ This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) as a base.
21
+
22
+ ### Models Merged
23
+
24
+ The following models were included in the merge:
25
+ * [cognitivecomputations/dolphin-2.2.1-mistral-7b](https://huggingface.co/cognitivecomputations/dolphin-2.2.1-mistral-7b)
26
+ * [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta)
27
+ * [NousResearch/Hermes-2-Pro-Mistral-7B](https://huggingface.co/NousResearch/Hermes-2-Pro-Mistral-7B)
28
+
29
+ ### Configuration
30
+
31
+ The following YAML configuration was used to produce this model:
32
+
33
+ ```yaml
34
+ models:
35
+ - model: cognitivecomputations/dolphin-2.2.1-mistral-7b
36
+ parameters:
37
+ density: [1, 0.7, 0.1] # density gradient
38
+ weight: 1.0
39
+ - model: HuggingFaceH4/zephyr-7b-beta
40
+ parameters:
41
+ density: 0.5
42
+ weight: [0, 0.3, 0.7, 1] # weight gradient
43
+ - model: NousResearch/Hermes-2-Pro-Mistral-7B
44
+ parameters:
45
+ density: 0.33
46
+ weight:
47
+ - filter: mlp
48
+ value: 0.5
49
+ - value: 0
50
+ merge_method: dare_ties
51
+ base_model: mistralai/Mistral-7B-v0.1
52
+ parameters:
53
+ normalize: true
54
+ int8_mask: true
55
+ dtype: float16
56
+ ```