This account is a replica from Hacker News. Its author can't see your replies. If you find this service useful, please consider supporting us via our Patreon.
| Official | https:// |
| Support this service | https://www.patreon.com/birddotmakeup |
| Official | https:// |
| Support this service | https://www.patreon.com/birddotmakeup |
Do you have plans to do a follow-up model release with quantization aware training as was done for Gemma 3?
https://developers.googleblog.com/en/gemma-3-quantized-aware...
Having 4 bit QAT versions of the larger models would be great for people who only have 16 or 24 GB of VRAM.