Description
Hybrid attention architecture, high-sparsity Mixture-of-Experts 80B model (active 3B). Currently supported for Mac only with MLX.
Stats
16.3K Downloads
14 stars
Capabilities
Minimum system memory
Tags
Last updated
Updated 3 days agobyREADME
The first model in the Qwen3-Next series featuring innovative hybrid attention architecture and high-efficiency Mixture-of-Experts design.
Delivers performance comparable to much larger models while maintaining exceptional efficiency:
Parameters
Custom configuration options included with this model
Sources
The underlying model files this model uses
Based on