Nvidia's new Nemotron 3 blends the Mamba hybrid sequence model with 31.6 B parameters, keeping only 3 B active per step. The design outperforms gpt‑oss‑20B and rivals Qwen3‑30B, while integrating LatentMoE and the Artificial Analysis Index. Open‑source fans, see how this hybrid architecture pushes efficiency and scaling. #Nvidia #Nemotron3 #MambaHybrid #LatentMoE
🔗 https://aidailypost.com/news/nvidias-nemotron-3-uses-mamba-hybrid-316b-params-3b-active-per-step
