Improve README.md
Browse files
README.md
CHANGED
|
@@ -2,4 +2,25 @@
|
|
| 2 |
license: other
|
| 3 |
license_name: tongyi-qianwen
|
| 4 |
license_link: https://huggingface.co/Qwen/Qwen1.5-32B/blob/main/LICENSE
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 5 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
license: other
|
| 3 |
license_name: tongyi-qianwen
|
| 4 |
license_link: https://huggingface.co/Qwen/Qwen1.5-32B/blob/main/LICENSE
|
| 5 |
+
base_model: Qwen/Qwen1.5-32B-Chat
|
| 6 |
+
tags:
|
| 7 |
+
- qwen
|
| 8 |
+
- chat
|
| 9 |
+
model-index:
|
| 10 |
+
- name: Qwen1.5-32B-Chat-iMat-GGUF
|
| 11 |
+
results: []
|
| 12 |
+
language:
|
| 13 |
+
- en
|
| 14 |
---
|
| 15 |
+
|
| 16 |
+
# Qwen1.5-32B-Chat-iMat-GGUF
|
| 17 |
+
|
| 18 |
+
Source Model: [Qwen/Qwen1.5-32B-Chat](https://huggingface.co/Qwen/Qwen1.5-32B-Chat)
|
| 19 |
+
|
| 20 |
+
Quantized with [llama.cpp](https://github.com/ggerganov/llama.cpp) commit [46acb3676718b983157058aecf729a2064fc7d34](https://github.com/ggerganov/llama.cpp/commit/46acb3676718b983157058aecf729a2064fc7d34)
|
| 21 |
+
|
| 22 |
+
Imatrix was generated from the f16 gguf via this command:
|
| 23 |
+
|
| 24 |
+
./imatrix -c 512 -m $out_path/$base_quant_name -f $llama_cpp_path/groups_merged.txt -o $out_path/imat-f16-gmerged.dat
|
| 25 |
+
|
| 26 |
+
Using the dataset from [here](https://github.com/ggerganov/llama.cpp/discussions/5263#discussioncomment-8395384)
|