Upload folder using huggingface_hub (#1)
Browse files- 73ec4d6ae0ba89b950ffe8cf46e003d0074ec191f2db34521ad2deda2243d854 (3a905f7416287f90f0a9f8f0af8f2d5918aa0b2e)
- 043e6100ab09610cb76d2c43ed01e9db259218dd199aec990f500ee41a292ffa (205e7530f2cf732ffc5a53bb2c578521f8d7357c)
- 83e73de9044aecab9d05413793fc0425828269f030fae638dcacacd604a2ae1c (e3c56a46f659b7f5d88dd2b570ef3c2cb5422543)
- eef4f1b58b63c00996d87396de9952c46b1a5d6440dbacaa56c77a8f74b7a5a5 (2aa2a6a2c6e4f91ee34684ffea39e1b9f8fd1096)
- c03c9cc8589e77302d38cf391241c3154f877116ecf9ecca78d75c068691d516 (792e9c44ff1547c8475584a1ac6b24fa04d44cdb)
- c5419543c7dc001d35439d9cd90d3fb39620a16843478a916c6e7d45269bb1ed (ccc19e384ff589c7d160296782c1f063bbbfa15c)
- 0b14ba5556def7950fcee4444ebc17e29dfbd3d46c8b4a636c59c4fa031e289f (2ba069960790cee8f231df242d1bab1ebbc4b914)
- 27b9c4e22a1cdc3d00f6f6684fe72ed56309414885337be3be69b6fb925061d2 (784d475209375848d429a3d7879b8746180dbd62)
- 059ce3fe69c1cc4270823951bc41ce7226b2d90d82d2aac2872298a380df9dce (082a790928fd9812a0602d79f6ab9069096c1cf3)
- f0c8da21150162135d497bbdc0751f9fcfa9724d000856e1f61f134d57400b6a (6b3f6dbcd0c13fe19a2eaf53860df29aa3e39fc8)
- 2a582eb541d806df1b61779ab69098b912adae85e263d2065a3b2d4fd93f9961 (e4548859a44f5bea440317c9b848070abd7326e0)
- b1288eb1eda04d784d8a4b0e32461a6d1561ef5e91cc66a8fd9cca1264b88b5e (dd7c0aef36ba2678e19a58eb1dcdf1c5e52ba458)
- 50dcb42f588bf2f9afb27e79d6ea1a2d274461e1c1177a30c3544dcc0561283c (6e19d2a9125c34f35f3335dbbf7d55a275449c6a)
- 725b16ef6994318f1702d4706f425cde115a1825dc62a698866bf61ae8ded98a (de68f26026076286a7fb073fb67c760f2744256f)
- 35eaaf2dfafa2c8fa4a0cba74d9e93425dfead29f25be45727a3e1cf7ff255f5 (2f1f7d01bd16212be1a3e5d3a6970d3fbec6eadb)
- .gitattributes +17 -0
- README.md +46 -0
- llama-3.2-Korean-Bllossom-3B-GGUF_imatrix.dat +3 -0
- llama-3.2-Korean-Bllossom-3B.IQ1_M.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.IQ1_S.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.IQ2_XS.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.IQ3_XS.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.IQ4_XS.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q2_K.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q3_K_L.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q3_K_M.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q3_K_S.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q4_K_M.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q4_K_S.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q5_K_M.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q5_K_S.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q6_K.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.Q8_0.gguf +3 -0
- llama-3.2-Korean-Bllossom-3B.fp16.gguf +3 -0
|
@@ -33,3 +33,20 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
| 34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
| 35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
| 36 |
+
llama-3.2-Korean-Bllossom-3B-GGUF_imatrix.dat filter=lfs diff=lfs merge=lfs -text
|
| 37 |
+
llama-3.2-Korean-Bllossom-3B.IQ1_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 38 |
+
llama-3.2-Korean-Bllossom-3B.IQ1_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 39 |
+
llama-3.2-Korean-Bllossom-3B.IQ2_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
| 40 |
+
llama-3.2-Korean-Bllossom-3B.IQ3_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
| 41 |
+
llama-3.2-Korean-Bllossom-3B.IQ4_XS.gguf filter=lfs diff=lfs merge=lfs -text
|
| 42 |
+
llama-3.2-Korean-Bllossom-3B.Q2_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 43 |
+
llama-3.2-Korean-Bllossom-3B.Q3_K_L.gguf filter=lfs diff=lfs merge=lfs -text
|
| 44 |
+
llama-3.2-Korean-Bllossom-3B.Q3_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 45 |
+
llama-3.2-Korean-Bllossom-3B.Q3_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 46 |
+
llama-3.2-Korean-Bllossom-3B.Q4_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 47 |
+
llama-3.2-Korean-Bllossom-3B.Q5_K_M.gguf filter=lfs diff=lfs merge=lfs -text
|
| 48 |
+
llama-3.2-Korean-Bllossom-3B.Q5_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
| 49 |
+
llama-3.2-Korean-Bllossom-3B.Q6_K.gguf filter=lfs diff=lfs merge=lfs -text
|
| 50 |
+
llama-3.2-Korean-Bllossom-3B.Q8_0.gguf filter=lfs diff=lfs merge=lfs -text
|
| 51 |
+
llama-3.2-Korean-Bllossom-3B.fp16.gguf filter=lfs diff=lfs merge=lfs -text
|
| 52 |
+
llama-3.2-Korean-Bllossom-3B.Q4_K_S.gguf filter=lfs diff=lfs merge=lfs -text
|
|
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
---
|
| 2 |
+
tags:
|
| 3 |
+
- quantized
|
| 4 |
+
- 2-bit
|
| 5 |
+
- 3-bit
|
| 6 |
+
- 4-bit
|
| 7 |
+
- 5-bit
|
| 8 |
+
- 6-bit
|
| 9 |
+
- 8-bit
|
| 10 |
+
- GGUF
|
| 11 |
+
- text-generation
|
| 12 |
+
- text-generation
|
| 13 |
+
model_name: llama-3.2-Korean-Bllossom-3B-GGUF
|
| 14 |
+
base_model: Bllossom/llama-3.2-Korean-Bllossom-3B
|
| 15 |
+
inference: false
|
| 16 |
+
model_creator: Bllossom
|
| 17 |
+
pipeline_tag: text-generation
|
| 18 |
+
quantized_by: MaziyarPanahi
|
| 19 |
+
---
|
| 20 |
+
# [MaziyarPanahi/llama-3.2-Korean-Bllossom-3B-GGUF](https://huggingface.co/MaziyarPanahi/llama-3.2-Korean-Bllossom-3B-GGUF)
|
| 21 |
+
- Model creator: [Bllossom](https://huggingface.co/Bllossom)
|
| 22 |
+
- Original model: [Bllossom/llama-3.2-Korean-Bllossom-3B](https://huggingface.co/Bllossom/llama-3.2-Korean-Bllossom-3B)
|
| 23 |
+
|
| 24 |
+
## Description
|
| 25 |
+
[MaziyarPanahi/llama-3.2-Korean-Bllossom-3B-GGUF](https://huggingface.co/MaziyarPanahi/llama-3.2-Korean-Bllossom-3B-GGUF) contains GGUF format model files for [Bllossom/llama-3.2-Korean-Bllossom-3B](https://huggingface.co/Bllossom/llama-3.2-Korean-Bllossom-3B).
|
| 26 |
+
|
| 27 |
+
### About GGUF
|
| 28 |
+
|
| 29 |
+
GGUF is a new format introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
|
| 30 |
+
|
| 31 |
+
Here is an incomplete list of clients and libraries that are known to support GGUF:
|
| 32 |
+
|
| 33 |
+
* [llama.cpp](https://github.com/ggerganov/llama.cpp). The source project for GGUF. Offers a CLI and a server option.
|
| 34 |
+
* [llama-cpp-python](https://github.com/abetlen/llama-cpp-python), a Python library with GPU accel, LangChain support, and OpenAI-compatible API server.
|
| 35 |
+
* [LM Studio](https://lmstudio.ai/), an easy-to-use and powerful local GUI for Windows and macOS (Silicon), with GPU acceleration. Linux available, in beta as of 27/11/2023.
|
| 36 |
+
* [text-generation-webui](https://github.com/oobabooga/text-generation-webui), the most widely used web UI, with many features and powerful extensions. Supports GPU acceleration.
|
| 37 |
+
* [KoboldCpp](https://github.com/LostRuins/koboldcpp), a fully featured web UI, with GPU accel across all platforms and GPU architectures. Especially good for story telling.
|
| 38 |
+
* [GPT4All](https://gpt4all.io/index.html), a free and open source local running GUI, supporting Windows, Linux and macOS with full GPU accel.
|
| 39 |
+
* [LoLLMS Web UI](https://github.com/ParisNeo/lollms-webui), a great web UI with many interesting and unique features, including a full model library for easy model selection.
|
| 40 |
+
* [Faraday.dev](https://faraday.dev/), an attractive and easy to use character-based chat GUI for Windows and macOS (both Silicon and Intel), with GPU acceleration.
|
| 41 |
+
* [candle](https://github.com/huggingface/candle), a Rust ML framework with a focus on performance, including GPU support, and ease of use.
|
| 42 |
+
* [ctransformers](https://github.com/marella/ctransformers), a Python library with GPU accel, LangChain support, and OpenAI-compatible AI server. Note, as of time of writing (November 27th 2023), ctransformers has not been updated in a long time and does not support many recent models.
|
| 43 |
+
|
| 44 |
+
## Special thanks
|
| 45 |
+
|
| 46 |
+
๐ Special thanks to [Georgi Gerganov](https://github.com/ggerganov) and the whole team working on [llama.cpp](https://github.com/ggerganov/llama.cpp/) for making all of this possible.
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:af09aec4546bf127e32c1483047fe81ae403dbe93681124bcb6138942fdcaebe
|
| 3 |
+
size 2988366
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6fee602678d40191ee550a2c8de65c282dbede35dff05b85e639e23b1815120d
|
| 3 |
+
size 924704096
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:23ea34217e48019b81f63340346e92368e40b15ac7736856767be49bb804830c
|
| 3 |
+
size 868670816
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:68eb78ac4d60c40ea702b0fb0f2b45287e900c13bff618b8c08fc9a151cb26b5
|
| 3 |
+
size 1101061472
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:60c8d2e41433f974e3b1873e4cd96eb6eff6d0385137356e7b0fbc3725bbcc1f
|
| 3 |
+
size 1477301600
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d444f1d1c9c7e59cf89fba683c2eb1564b2d3e57ef860b832b98bd9ce7009818
|
| 3 |
+
size 1829623136
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:8f5b28f5492e609391ec9b0b4f9fa54333c8b13329f5d17a399292228d0e1e19
|
| 3 |
+
size 1364448608
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:c1777202b0ddc90603a4ea78d5abb4651fcbaf213bdcd599f5b8d39043433b41
|
| 3 |
+
size 1815860576
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:4db95145c8e784e1c8af0a3a51658c9dfc5b61bf5fc85b801348150df0a1f611
|
| 3 |
+
size 1687672160
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1691d5372b58e3ff7e08f4a352ab52e9b8653dc6858334738f68d8f69fb4e836
|
| 3 |
+
size 1543361888
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:5e68db76d753feef10862bb461288a17e986498555fecbc394f2386c937b889e
|
| 3 |
+
size 2019890528
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:eeccb514b0682325b81665dbf3c1036354ba172205bac49094a4142f2e16dc5b
|
| 3 |
+
size 1928713568
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:92d6904ee424ae51f4a617fc3887878f9abc65ccc83a3070c239af10478321a0
|
| 3 |
+
size 2322666848
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:08414fe4d3d10fca3c9255dbe2ecd69d2a4d09328c63f0629abc9da7b13e7ceb
|
| 3 |
+
size 2270025056
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:6e505b446f243e49b3a3ee3a7eb66aee746ae2a85b37fbb4249d2e76c4dd80e2
|
| 3 |
+
size 2644366688
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:a7cd3b0716312c18de9bafc3ce9a295fb93e628c4d94e7e06ffb34819572ed17
|
| 3 |
+
size 3422412128
|
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:9383f9860f3292db632e3db0fbced42d5984beed4335cc006a38194c9a4e05d0
|
| 3 |
+
size 6434200640
|