sinjab commited on
Commit
ad61c51
·
verified ·
1 Parent(s): cb6ba6e

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +67 -0
README.md ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ license: apache-2.0
5
+ library_name: gguf
6
+ tags:
7
+ - reranker
8
+ - gguf
9
+ - llama.cpp
10
+ base_model: mixedbread-ai/mxbai-rerank-large-v2
11
+ ---
12
+
13
+ # mxbai-rerank-large-v2-F16-GGUF
14
+
15
+ This model was converted to GGUF format from [mixedbread-ai/mxbai-rerank-large-v2](https://huggingface.co/mixedbread-ai/mxbai-rerank-large-v2) using llama.cpp via the ggml.ai's GGUF-my-repo space.
16
+
17
+ Refer to the [original model card](https://huggingface.co/mixedbread-ai/mxbai-rerank-large-v2) for more details on the model.
18
+
19
+ ## Model Information
20
+
21
+ - **Base Model**: [mixedbread-ai/mxbai-rerank-large-v2](https://huggingface.co/mixedbread-ai/mxbai-rerank-large-v2)
22
+ - **Quantization**: F16
23
+ - **Format**: GGUF (GPT-Generated Unified Format)
24
+ - **Converted with**: llama.cpp
25
+
26
+ ## Quantization Details
27
+
28
+ This is a **F16** quantization of the original model:
29
+
30
+ - **F16**: Full 16-bit floating point - highest quality, largest size
31
+ - **Q8_0**: 8-bit quantization - high quality, good balance
32
+ - **Q4_K_M**: 4-bit quantization with medium quality - smaller size, faster inference
33
+
34
+ ## Usage
35
+
36
+ This model can be used with llama.cpp and other GGUF-compatible inference engines.
37
+
38
+ ```bash
39
+ # Example using llama.cpp
40
+ ./llama-rerank -m mxbai-rerank-large-v2-F16.gguf
41
+ ```
42
+
43
+ ## Model Files
44
+
45
+ | Quantization | Use Case |
46
+ |-------------|----------|
47
+ | F16 | Maximum quality, largest size |
48
+ | Q8_0 | High quality, good balance of size/performance |
49
+ | Q4_K_M | Good quality, smallest size, fastest inference |
50
+
51
+ ## Citation
52
+
53
+ If you use this model, please cite the original model:
54
+
55
+ ```bibtex
56
+ # See original model card for citation information
57
+ ```
58
+
59
+ ## License
60
+
61
+ This model inherits the license from the original model. Please refer to the [original model card](https://huggingface.co/mixedbread-ai/mxbai-rerank-large-v2) for license details.
62
+
63
+ ## Acknowledgements
64
+
65
+ - Original model by the authors of [mixedbread-ai/mxbai-rerank-large-v2](https://huggingface.co/mixedbread-ai/mxbai-rerank-large-v2)
66
+ - GGUF conversion via llama.cpp by ggml.ai
67
+ - Converted and uploaded by [sinjab](https://huggingface.co/sinjab)