Update README.md
Browse files
README.md
CHANGED
|
@@ -51,10 +51,11 @@ Using [llama.cpp quantize cae9fb4](https://github.com/ggerganov/llama.cpp/commit
|
|
| 51 |
|
| 52 |
- Bravo IQ1_S worse than Alpha?
|
| 53 |
- [Latent loss](https://huggingface.co/Eviation/flux-imatrix/blob/main/images/latent_loss.png)
|
| 54 |
-
- [Per layer quantization cost](https://huggingface.co/Eviation/flux-imatrix/blob/main/images/
|
| 55 |
-
|
| 56 |
|
|
|
|
| 57 |
| Weight name |
|
|
|
|
| 58 |
| double_blocks.%d.txt_attn.proj.weight |
|
| 59 |
| double_blocks.%d.txt_mod.lin.weight |
|
| 60 |
| time_in.in_layer.weight && time_in.out_layer.weight |
|
|
|
|
| 51 |
|
| 52 |
- Bravo IQ1_S worse than Alpha?
|
| 53 |
- [Latent loss](https://huggingface.co/Eviation/flux-imatrix/blob/main/images/latent_loss.png)
|
| 54 |
+
- [Per layer quantization cost](https://huggingface.co/Eviation/flux-imatrix/blob/main/images/casting_cost.png) from [chrisgoringe/casting_cost](https://github.com/chrisgoringe/mixed-gguf-converter/blob/main/costs/casting_cost.yaml)
|
|
|
|
| 55 |
|
| 56 |
+
### Per weight type quantization loss in increasing order:
|
| 57 |
| Weight name |
|
| 58 |
+
| ----------- |
|
| 59 |
| double_blocks.%d.txt_attn.proj.weight |
|
| 60 |
| double_blocks.%d.txt_mod.lin.weight |
|
| 61 |
| time_in.in_layer.weight && time_in.out_layer.weight |
|