google/gemma-3-4b-it-qat-q4_0-gguf
Image-Text-to-Text • 4B • Updated
• 16.6k • 241
Quantization Aware Trained (QAT) Gemma 3 checkpoints. The model preserves similar quality as half precision while using 3x less memory
Note ^ GGUFs to be used in llama.cpp and Ollama. We strongly recommend using the IT models.
Note ^ unquantized QAT-based checkpoints that allow quantizing while retaining similar quality to half precision