caiovicentino1/Qwopus-MoE-35B-A3B-HLWQ-Q5
Text Generation • 35B • Updated • 1.86k • 4
Massive MoE models ≥100B quantized with HLWQ · consumer deploy via vLLM expert offload
Note 35B · 128 × top-8 · 16 GB HLWQ Q5 (legacy name)
Note 30B MoE · consumer GPU via expert offload
Note 27B Gemma-4 MoE · 128 × top-8 · multimodal · 16.6 GB HLWQ Q5