A compact, quantized text model running in 8-bit precision via the MLX framework, optimized for Apple Silicon hardware. It trades some numerical precision for reduced memory footprint, making it practical to run locally on machines that might otherwise struggle with full-precision weights. Behavior stays close to the base Gemma 4 family, with the quantization introducing minimal but present degradation.