gemma-4
Collection
in mxfp4, mxfp8, and Deckard(qx) • 26 items • Updated
Brainwaves
arc arc/e boolq hswag obkqa piqa wino
bf16 0.490,0.674,0.793,0.612,0.416,0.756,0.669
mxfp8 0.480,0.656,0.797,0.608,0.400,0.755,0.665
mxfp4 0.455,0.607,0.851,0.585,0.402,0.744,0.651
Quant Perplexity Peak Memory Tokens/sec
mxfp8 35.937 ± 0.525 14.80 GB 1153
mxfp4 36.746 ± 0.534 11.06 GB 1030
The model chat template has been updated with the latest Jinja from Google
See parent model for instructions on install and use with Transformers.
-G
4-bit
Base model
google/gemma-4-E4B-it