Qwen3-4B-Instruct-2507_w8a8_g128

  • My quant for Rock5B (RK3588 board)
  • Author: @flylcw
  • 建议运行环境:RKLLM-Toolkit在1.2.x应该都可运行,NPU驱动版本需>v0.9.8

Conversion details:

  • RKLLM-Toolkit version: v1.2.2
  • NPU driver: v0.9.8
  • Python: 3.11
  • Quantization: w8a8_g128
  • Output: single-file .rkllm artifact
  • Modifications: quantization (w8a8_g128), export to .rkllm format for RK3588 SBCs
Downloads last month
7
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for flylcw/Qwen3-4B-Instruct-2507_w8a8_g128_1.2.2_rkllm

Finetuned
(167)
this model