geoffmunn commited on
Commit
fd146d7
·
verified ·
1 Parent(s): de03cbb

Models highlighted

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -34,7 +34,7 @@ These variants were built from a **f16** base model to ensure consistency across
34
 
35
  | Level | Speed | Size | Recommendation |
36
  |-----------|-----------|------------|--------------------------------------------------------------------|
37
- | Q2_K | ⚡ Fastest | 347 MB | **DO NOT USE.** Could not provide an answer to any question. |
38
  | Q3_K_S | ⚡ Fast | 390 MB | Not recommended, did not appear in any top 3 results. |
39
  | Q3_K_M | ⚡ Fast | 414 MB | First place in the bat & ball question, no other top 3 appearances.|
40
  | Q4_K_S | 🚀 Fast | 471 MB | A good option for technical, low-temperature questions. |
@@ -61,7 +61,7 @@ It’s ideal for:
61
  ## Model anaysis and rankings
62
 
63
  I have run each of these models across 6 questions, and ranked them all based on the quality of the anwsers.
64
- Qwen3-0.6B-f16:Q5_K_M is the best model across all question types, but if you want to play it safe with a higher precision model, then you could consider using Qwen3-0.6B:Q8_0.
65
 
66
  You can read the results here: [Qwen3-0.6b-analysis.md](Qwen3-0.6b-analysis.md)
67
 
 
34
 
35
  | Level | Speed | Size | Recommendation |
36
  |-----------|-----------|------------|--------------------------------------------------------------------|
37
+ | Q2_K | ⚡ Fastest | 347 MB | 🚨 **DO NOT USE.** Could not provide an answer to any question. |
38
  | Q3_K_S | ⚡ Fast | 390 MB | Not recommended, did not appear in any top 3 results. |
39
  | Q3_K_M | ⚡ Fast | 414 MB | First place in the bat & ball question, no other top 3 appearances.|
40
  | Q4_K_S | 🚀 Fast | 471 MB | A good option for technical, low-temperature questions. |
 
61
  ## Model anaysis and rankings
62
 
63
  I have run each of these models across 6 questions, and ranked them all based on the quality of the anwsers.
64
+ **Qwen3-0.6B-f16:Q5_K_M** is the best model across all question types, but if you want to play it safe with a higher precision model, then you could consider using **Qwen3-0.6B:Q8_0**.
65
 
66
  You can read the results here: [Qwen3-0.6b-analysis.md](Qwen3-0.6b-analysis.md)
67