Update README.md
Browse files
README.md
CHANGED
|
@@ -17,7 +17,7 @@ Here's a precise analysis of YOYO-V2-dwq5's performance compared to the other qu
|
|
| 17 |
|
| 18 |
Comparison Table (YOYO-V2 Quantized Variants)
|
| 19 |
```bash
|
| 20 |
-
Task
|
| 21 |
arc_challenge 0.523 0.511 0.497 0.532
|
| 22 |
arc_easy 0.682 0.655 0.657 0.685
|
| 23 |
boolq 0.883 0.879 0.876 0.886
|
|
@@ -31,8 +31,8 @@ YOYO-V2-q6 scores are highest across all tasks in this dataset.
|
|
| 31 |
|
| 32 |
π Critical Insights from YOYO-V2's Internal Quantization Comparison
|
| 33 |
|
| 34 |
-
```bash
|
| 35 |
YOYO-V2-dwq5 Consistently Improves Over Lower-DWQ Variants
|
|
|
|
| 36 |
DWQ5 surpasses dwq4 in all tasks (e.g., +0.002 on arc_easy, +0.007 on boolq).
|
| 37 |
DWQ5 surpasses dwq3 in all tasks (e.g., +0.026 on arc_easy, +0.014 on boolq).
|
| 38 |
```
|
|
@@ -62,16 +62,16 @@ piqa: q6βs lead (0.782 vs dwq5βs 0.778) is 1.3% β critical for logic reas
|
|
| 62 |
|
| 63 |
π― Practical Takeaways for Model Selection
|
| 64 |
```bash
|
| 65 |
-
|
| 66 |
-
|
| 67 |
-
|
| 68 |
```
|
| 69 |
|
| 70 |
-
For most use cases,
|
| 71 |
|
| 72 |
-
|
| 73 |
|
| 74 |
-
|
| 75 |
|
| 76 |
This confirms that YOYO-V2βs performance steadily improves with higher quantization fidelity within its own variants, but the fixed Q6 quantization still delivers edge gains for critical tasks where minor precision losses are unacceptable.
|
| 77 |
|
|
|
|
| 17 |
|
| 18 |
Comparison Table (YOYO-V2 Quantized Variants)
|
| 19 |
```bash
|
| 20 |
+
Task dwq5 dwq4 dwq3 q6
|
| 21 |
arc_challenge 0.523 0.511 0.497 0.532
|
| 22 |
arc_easy 0.682 0.655 0.657 0.685
|
| 23 |
boolq 0.883 0.879 0.876 0.886
|
|
|
|
| 31 |
|
| 32 |
π Critical Insights from YOYO-V2's Internal Quantization Comparison
|
| 33 |
|
|
|
|
| 34 |
YOYO-V2-dwq5 Consistently Improves Over Lower-DWQ Variants
|
| 35 |
+
```bash
|
| 36 |
DWQ5 surpasses dwq4 in all tasks (e.g., +0.002 on arc_easy, +0.007 on boolq).
|
| 37 |
DWQ5 surpasses dwq3 in all tasks (e.g., +0.026 on arc_easy, +0.014 on boolq).
|
| 38 |
```
|
|
|
|
| 62 |
|
| 63 |
π― Practical Takeaways for Model Selection
|
| 64 |
```bash
|
| 65 |
+
Quant Best For Why
|
| 66 |
+
dwq5 Hardware with moderate resources Best balance between speed and accuracy (e.g., 5-bit DWQ)
|
| 67 |
+
q6 High-precision tasks (e.g., reasoning) Slightly better than dwq5 in 4+ tasks; optimal for stability
|
| 68 |
```
|
| 69 |
|
| 70 |
+
For most use cases, q6 is still the top performer (1.3β2.0% edge over dwq5 in tasks like boolq and piqa).
|
| 71 |
|
| 72 |
+
dwq5 is ideal if you need to reduce memory footprint while still achieving near-q6 performance (e.g., in edge devices).
|
| 73 |
|
| 74 |
+
dwq5 outperforms the lower-DWQ quantizations (dwq3, dwq4) across all tasks, showing a clear progression in precision as the DWQ bitwidth increases from 3 β 5 bits. However, it does not surpass YOYO-V2-q6 β instead, q6 maintains a small but consistent lead (0.005β0.013) in high-precision tasks like boolq and piqa.
|
| 75 |
|
| 76 |
This confirms that YOYO-V2βs performance steadily improves with higher quantization fidelity within its own variants, but the fixed Q6 quantization still delivers edge gains for critical tasks where minor precision losses are unacceptable.
|
| 77 |
|