A compact, quantized text model running at 6-bit precision via the MLX framework, optimized for local inference on Apple Silicon hardware. Its reduced bit-depth means a smaller memory footprint compared to full-precision variants, with some trade-off in output fidelity. It behaves like a lightweight workhorse suited for on-device experimentation.