A compact, quantized text model optimized for local inference on Apple Silicon hardware via the MLX framework. The 4-bit quantization keeps memory footprint small, making it practical for running on consumer machines, though this compression introduces some quality trade-offs compared to full-precision versions. It handles everyday text tasks in a resource-conscious way.