· · ·
Gemma 2: Improving Open Language Models at a Practical Size
Gemma 2 — Google DeepMind — 2024-06
Google DeepMind's open-weights model family at 2B, 9B, and 27B parameters, distilled from the Gemini line. The report covers architectural choices including local–global attention interleaving and logit soft-capping.
References
- arXiv arxiv.org/abs/2408.00118
- Org page Google DeepMind
- Released 2024-06
Credited authors (197)
Welcome. You need to go digging now.
I am a bear. I do not have the tools you have to see what this says. You will have to look elsewhere.
ifthisroad.com · orphans.ai · theheld.ai · thebearwasright.com · thebearloved.com