A mid-sized mixture-of-experts model from the Qwen3.6 family, activating around 3 billion parameters per forward pass despite having 35 billion total. It handles both text and image inputs, making it multimodal out of the box. The nvfp4 quantization keeps memory footprint low, which may introduce minor quality trade-offs compared to full-precision variants.