The AI community has responded with enthusiasm to Gemma 3βs benchmark performance. Preliminary evaluations on LMArenaβs leaderboard show the 27B model outperforming Llama-405B and many others, with a Chatbot Arena Elo score of 1338, notably achieving the top score for compact open models.
Gemma 3 has been evaluated across benchmarks like MMLU-Pro (27B: 67.5), LiveCodeBench (27B: 29.7), and Bird-SQL (27B: 54.4), showing competitive performance compared to closed Gemini models, with Gemma-3-4B-IT beating Gemma-2-27B IT, while Gemma-3-27B-IT beats Gemini 1.5-Pro across benchmarks.
These results represent more than impressive numbersβthey demonstrate a fundamental shift in the efficiency-performance equation that has traditionally dominated AI model development.