A mid-sized open-weight model from Google's Gemma 4 family, quantized to FP8 precision by protoLabsAI for more efficient deployment. The FP8 quantization means it runs faster and uses less memory than the full-precision version, with a modest trade-off in output fidelity. It handles text tasks competently and suits environments where hardware constraints matter.