A mid-sized mixture-of-experts base model that activates only 3 billion of its 35 billion parameters per forward pass, making it computationally lean relative to its total capacity. As a base model, it arrives without instruction tuning or chat formatting, meaning it excels at completion tasks and serves as a foundation for fine-tuning rather than direct conversation.