Eviation commited on
Commit
b0f9180
·
verified ·
1 Parent(s): db1157d

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -51,10 +51,11 @@ Using [llama.cpp quantize cae9fb4](https://github.com/ggerganov/llama.cpp/commit
51
 
52
  - Bravo IQ1_S worse than Alpha?
53
  - [Latent loss](https://huggingface.co/Eviation/flux-imatrix/blob/main/images/latent_loss.png)
54
- - [Per layer quantization cost](https://huggingface.co/Eviation/flux-imatrix/blob/main/images/latent_loss.png) from [chrisgoringe/casting_cost](https://github.com/chrisgoringe/mixed-gguf-converter/blob/main/costs/casting_cost.yaml)
55
-
56
 
 
57
  | Weight name |
 
58
  | double_blocks.%d.txt_attn.proj.weight |
59
  | double_blocks.%d.txt_mod.lin.weight |
60
  | time_in.in_layer.weight && time_in.out_layer.weight |
 
51
 
52
  - Bravo IQ1_S worse than Alpha?
53
  - [Latent loss](https://huggingface.co/Eviation/flux-imatrix/blob/main/images/latent_loss.png)
54
+ - [Per layer quantization cost](https://huggingface.co/Eviation/flux-imatrix/blob/main/images/casting_cost.png) from [chrisgoringe/casting_cost](https://github.com/chrisgoringe/mixed-gguf-converter/blob/main/costs/casting_cost.yaml)
 
55
 
56
+ ### Per weight type quantization loss in increasing order:
57
  | Weight name |
58
+ | ----------- |
59
  | double_blocks.%d.txt_attn.proj.weight |
60
  | double_blocks.%d.txt_mod.lin.weight |
61
  | time_in.in_layer.weight && time_in.out_layer.weight |