Gemma 4 has a systemic attention failure. Here’s the proof.
I've spent months building a diagnostic method for large language models. It catches what standard benchmarks miss – distributional collapse inside tensors, not just loss or perplexity. Gemma 4 26B A4B fails it. I analyzed Gemma 4 26B A4B (Q8_0) qu…