A large mixture-of-experts model that activates roughly 10 billion parameters per forward pass despite having 122 billion total, keeping inference costs manageable while drawing on a wide pool of specialized knowledge. It handles complex reasoning, coding, and multilingual tasks with the fluency you'd expect from a much heavier model. The NVFP4 quantization means it trades a small amount of numerical precision for significantly reduced memory footprint.