A compact, open-weight model from Google's Gemma 4 family, quantized to 8-bit precision for efficient local inference via MLX. The 8-bit quantization keeps memory footprint manageable while preserving much of the original model's capability, though some precision loss is expected compared to full-weight versions. It handles text-in, text-out tasks and runs well on Apple Silicon hardware through the MLX format.