Core Insights - Google has officially announced the launch of Gemini 3 Flash, the fastest and most cost-effective model in the Gemini 3 series, which outperforms flagship models in certain performance metrics while being cheaper and faster [1][3]. Performance and Efficiency - Gemini 3 Flash has surpassed the previous flagship model 2.5 Pro in performance and efficiency, achieving a speed increase of 3 times while being offered at a significantly lower price [3]. - In benchmark tests, Gemini 3 Flash scored 78% in SWE-bench Verified, outperforming both Gemini 3 Pro and Claude Sonnet 4.5, and achieved 81.2% in the MMMU-Pro benchmark, exceeding GPT-5.2 and Claude Sonnet 4.5 by several percentage points [4][5]. Cost-Effectiveness - The input cost for Gemini 3 Flash is $0.50 per million tokens, and the output cost is $3 per million tokens, making it the most cost-effective model compared to Claude Sonnet 4.5 and GPT-5.2, which have output costs of $15 and $14 per million tokens, respectively [6]. - Developers have reported that switching from GPT-4 or Gemini 3 Pro to Gemini 3 Flash could reduce costs by 50%-70% [8]. Market Position and Adoption - Gemini 3 Flash is set to replace the previous 2.5 Flash model in the Gemini App, becoming the default model for all users, including free users, while Gemini 3 Pro remains available for more complex tasks [8][9]. - Since its release, the internal API for Gemini 3 has been processing over 1 trillion tokens daily, indicating strong market adoption and usage for various applications, including code simulation and interactive game design [8][9].
狙击Open AI!谷歌一个月内连发「数弹」