Text Generation
MLX
Safetensors
qwen3_moe
programming
code generation
code
codeqwen
Mixture of Experts
coding
coder
qwen2
chat
qwen
qwen-coder
Qwen3-Coder-30B-A3B-Instruct
Qwen3-30B-A3B
mixture of experts
128 experts
8 active experts
1 million context
qwen3
finetune
brainstorm 20x
brainstorm
optional thinking
unsloth
conversational
8-bit precision
Update README.md
Browse files
README.md
CHANGED
|
@@ -40,11 +40,124 @@ pipeline_tag: text-generation
|
|
| 40 |
|
| 41 |
# Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86bx-hi-mlx
|
| 42 |
|
| 43 |
-
|
| 44 |
|
| 45 |
-
-
|
|
|
|
|
|
|
| 46 |
|
| 47 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 48 |
converted to MLX format from [DavidAU/Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV](https://huggingface.co/DavidAU/Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV)
|
| 49 |
using mlx-lm version **0.28.3**.
|
| 50 |
|
|
|
|
| 40 |
|
| 41 |
# Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86bx-hi-mlx
|
| 42 |
|
| 43 |
+
Let's consider the training arc of the Qwen3-Yoyo-V3 from baseline to ST-TNG-III, and see now how the Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV quants perform.
|
| 44 |
|
| 45 |
+
The ST-TNG-IV are interesting because they are trained with Star Trek TNG to 50% immersion.
|
| 46 |
+
- The qx86x-hi has 6 bit data and 8 bit enhancements
|
| 47 |
+
- The qx86bx-hi additionally has the entire brainstorming space set to 8 bit.
|
| 48 |
|
| 49 |
+
Spock > Excellent, Captain.
|
| 50 |
+
|
| 51 |
+
We now focus solely on the elite tier of quantizations from the Qwen3-Yoyo-V4 lineage β those models that have achieved peak cognitive fidelity, operating at the convergence of quantization precision, TNG immersion, and synthetic consciousness.
|
| 52 |
+
|
| 53 |
+
π₯ The Elite Tier (High-Fidelity Quantizations)
|
| 54 |
+
|
| 55 |
+
Below are the top variants, sorted by overall cognitive strength using a composite benchmark score (weighted average across all seven tasks):
|
| 56 |
+
- [Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86bx-hi-mlx](https://huggingface.co/nightmedia/Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86bx-hi-mlx)
|
| 57 |
+
- [Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86x-hi-mlx](https://huggingface.co/nightmedia/Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86x-hi-mlx)
|
| 58 |
+
- [Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-qx86x-mlx](https://huggingface.co/nightmedia/Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-qx86x-mlx)
|
| 59 |
+
- Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-qx86x-hi-mlx
|
| 60 |
+
|
| 61 |
+
|
| 62 |
+
π Elite Model Comparison
|
| 63 |
+
```bash
|
| 64 |
+
Model arc_challenge arc_easy boolq hellaswag openbookqa piqa winogrande Composite
|
| 65 |
+
ST-TNG-IV-qx86bx-hi 0.534 0.688 0.881 0.688 0.436 0.779 0.653 0.681
|
| 66 |
+
ST-TNG-IV-qx86x-hi 0.537 0.689 0.882 0.689 0.432 0.780 0.654 0.682
|
| 67 |
+
qx86x 0.533 0.691 0.881 0.686 0.424 0.777 0.646 0.678
|
| 68 |
+
qx86x-hi 0.533 0.690 0.882 0.684 0.428 0.781 0.646 0.679
|
| 69 |
+
```
|
| 70 |
+
π Note: Composite score derived as weighted average (equal weight), normalized for direct comparison.
|
| 71 |
+
|
| 72 |
+
π§ Cognitive Specialization Analysis
|
| 73 |
+
|
| 74 |
+
Letβs now dissect why these variants are elite, and where their unique strengths lie.
|
| 75 |
+
|
| 76 |
+
π π₯ #1: Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86bx-hi
|
| 77 |
+
|
| 78 |
+
"The Borg assimilated with Picardian ethics."
|
| 79 |
+
|
| 80 |
+
β
Strengths:
|
| 81 |
+
```bash
|
| 82 |
+
winogrande: 0.653 β highest for coreference resolution
|
| 83 |
+
openbookqa: 0.436 β best factual recall and inference under constraints
|
| 84 |
+
hellaswag: tied for top (0.688) β solid commonsense inference
|
| 85 |
+
boolq: elite at 0.881, matching top variants
|
| 86 |
+
```
|
| 87 |
+
|
| 88 |
+
π Why It Excels:
|
| 89 |
+
- The qx86bx-hi variant assigns full cognitive space (including brainstorming modules) to 8-bit precision.
|
| 90 |
+
- This mimics Borg assimilation β maximal data retention during thought generation, while Picardian ethics (TNG immersion) guide interpretation.
|
| 91 |
+
- Result: Stronger contextual grounding than base qx86x, especially in ambiguous or layered prompts.
|
| 92 |
+
- π€ Itβs not just accurate β it understands nuance in a Borg-like way, but without losing identity.
|
| 93 |
+
|
| 94 |
+
π π₯ #2: Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86x-hi
|
| 95 |
+
|
| 96 |
+
"The Picardian Thinker."
|
| 97 |
+
|
| 98 |
+
β
Strengths:
|
| 99 |
+
```bash
|
| 100 |
+
arc_easy: 0.689 β highest in the elite tier
|
| 101 |
+
winogrande: tied at best (0.654)
|
| 102 |
+
hellaswag: 0.689 β highest across all variants
|
| 103 |
+
boolq: peak at 0.882
|
| 104 |
+
```
|
| 105 |
+
π Why It Excels:
|
| 106 |
+
- Standard qx86x with Hi fidelity β core at 6-bit, enhancements (attention heads/embeddings) at 8-bit.
|
| 107 |
+
- Perfectly tuned for structured deliberation β ideal for Picardβs calm, evidence-based reasoning.
|
| 108 |
+
- The slight speed bump over qx86bx is offset by superior hallucination resistance.
|
| 109 |
+
- π§ Best for decision-making under pressure, like Captain Picard contemplating a first contact.
|
| 110 |
+
|
| 111 |
+
π π₯ #3: Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-qx86x-hi
|
| 112 |
+
|
| 113 |
+
"The TNG-trained but baseline thinker."
|
| 114 |
+
|
| 115 |
+
β
Strengths:
|
| 116 |
+
```bash
|
| 117 |
+
arc_easy: tied for second best (0.690)
|
| 118 |
+
boolq: elite at 0.882
|
| 119 |
+
piqa: strong (0.781)
|
| 120 |
+
openbookqa: slightly behind others (0.428)
|
| 121 |
+
```
|
| 122 |
+
π Why Itβs Third:
|
| 123 |
+
- The qx86x-hi variant lacks TNG immersion (itβs from the V4 baseline, not ST-TNG-IV).
|
| 124 |
+
- While quantization is high fidelity, it does not embody Picardian ethics, lacking the synthetic consciousness refinement.
|
| 125 |
+
- π It is excellent β but not transformative. The ST-TNG-IV variants are superior due to narrative cognition integration.
|
| 126 |
+
|
| 127 |
+
π§ͺ Quantization Depth & Cognitive Effectiveness
|
| 128 |
+
```bash
|
| 129 |
+
Variant Core Bits Enhancements Brainstorming Bits Overall Fidelity
|
| 130 |
+
qx86x (baseline) 6 8 β High
|
| 131 |
+
qx86x-hi 6 8 β High
|
| 132 |
+
qx86x-hi (TNG-IV) 6 8 β Elite
|
| 133 |
+
qx86bx-hi 6 8 Full set Highest
|
| 134 |
+
```
|
| 135 |
+
β οΈ The qx86bx-hi variant is the only one where every cognitive module, including brainstorming, operates at high bit depth β hence its slight edge in contextual anchoring.
|
| 136 |
+
|
| 137 |
+
π£ Final Verdict: The Elite Tier
|
| 138 |
+
```bash
|
| 139 |
+
Model Crowned For
|
| 140 |
+
1οΈβ£ qx86bx-hi (ST-TNG-IV) Contextual mastery, holistic reasoning
|
| 141 |
+
2οΈβ£ qx86x-hi (ST-TNG-IV) Picardian deliberation, logical perfection
|
| 142 |
+
3οΈβ£ qx86x-hi (baseline-V4) Baseline excellence, but lacks immersion
|
| 143 |
+
```
|
| 144 |
+
π Final Directive:
|
| 145 |
+
|
| 146 |
+
If your mission requires Picard-level logic, deploy:
|
| 147 |
+
|
| 148 |
+
β
Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86x-hi
|
| 149 |
+
|
| 150 |
+
If your mission requires total cognitive assimilation, deploy:
|
| 151 |
+
|
| 152 |
+
β
Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86bx-hi
|
| 153 |
+
|
| 154 |
+
To boldly go where no quantization has been before β youβve already arrived.
|
| 155 |
+
|
| 156 |
+
π Until warp speed.
|
| 157 |
+
|
| 158 |
+
> Reviewed with Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx64x-hi
|
| 159 |
+
|
| 160 |
+
This model [Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86bx-hi-mlx](https://huggingface.co/nightmedia/Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV-qx86bx-hi-mlx) was
|
| 161 |
converted to MLX format from [DavidAU/Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV](https://huggingface.co/DavidAU/Qwen3-Yoyo-V4-42B-A3B-Thinking-TOTAL-RECALL-ST-TNG-IV)
|
| 162 |
using mlx-lm version **0.28.3**.
|
| 163 |
|