Models highlighted
Browse files
README.md
CHANGED
|
@@ -34,7 +34,7 @@ These variants were built from a **f16** base model to ensure consistency across
|
|
| 34 |
|
| 35 |
| Level | Speed | Size | Recommendation |
|
| 36 |
|-----------|-----------|------------|--------------------------------------------------------------------|
|
| 37 |
-
| Q2_K | ⚡ Fastest | 347 MB | **DO NOT USE.** Could not provide an answer to any question. |
|
| 38 |
| Q3_K_S | ⚡ Fast | 390 MB | Not recommended, did not appear in any top 3 results. |
|
| 39 |
| Q3_K_M | ⚡ Fast | 414 MB | First place in the bat & ball question, no other top 3 appearances.|
|
| 40 |
| Q4_K_S | 🚀 Fast | 471 MB | A good option for technical, low-temperature questions. |
|
|
@@ -61,7 +61,7 @@ It’s ideal for:
|
|
| 61 |
## Model anaysis and rankings
|
| 62 |
|
| 63 |
I have run each of these models across 6 questions, and ranked them all based on the quality of the anwsers.
|
| 64 |
-
Qwen3-0.6B-f16:Q5_K_M is the best model across all question types, but if you want to play it safe with a higher precision model, then you could consider using Qwen3-0.6B:Q8_0
|
| 65 |
|
| 66 |
You can read the results here: [Qwen3-0.6b-analysis.md](Qwen3-0.6b-analysis.md)
|
| 67 |
|
|
|
|
| 34 |
|
| 35 |
| Level | Speed | Size | Recommendation |
|
| 36 |
|-----------|-----------|------------|--------------------------------------------------------------------|
|
| 37 |
+
| Q2_K | ⚡ Fastest | 347 MB | 🚨 **DO NOT USE.** Could not provide an answer to any question. |
|
| 38 |
| Q3_K_S | ⚡ Fast | 390 MB | Not recommended, did not appear in any top 3 results. |
|
| 39 |
| Q3_K_M | ⚡ Fast | 414 MB | First place in the bat & ball question, no other top 3 appearances.|
|
| 40 |
| Q4_K_S | 🚀 Fast | 471 MB | A good option for technical, low-temperature questions. |
|
|
|
|
| 61 |
## Model anaysis and rankings
|
| 62 |
|
| 63 |
I have run each of these models across 6 questions, and ranked them all based on the quality of the anwsers.
|
| 64 |
+
**Qwen3-0.6B-f16:Q5_K_M** is the best model across all question types, but if you want to play it safe with a higher precision model, then you could consider using **Qwen3-0.6B:Q8_0**.
|
| 65 |
|
| 66 |
You can read the results here: [Qwen3-0.6b-analysis.md](Qwen3-0.6b-analysis.md)
|
| 67 |
|