Google Launches Gemini 3.1 Pro, Setting New AI Performance and Cost Standards

Google’s Gemini 3.1 Pro Sets New AI Performance Standards

Google has unveiled Gemini 3.1 Pro, the latest iteration in its line of advanced language models, marking a significant leap in artificial intelligence capabilities. This release underscores Google’s commitment to leading the AI industry by delivering models that excel in complex reasoning, coding, and scientific knowledge.

Unprecedented Benchmark Achievements

Gemini 3.1 Pro has demonstrated exceptional performance across a range of industry-standard benchmarks, solidifying its position at the forefront of AI development. Notably, the model achieved a 77.1% score on the ARC-AGI-2 benchmark, which assesses abstract reasoning abilities. This score more than doubles that of its predecessor, Gemini 3 Pro, which scored 31.1%, and surpasses competitors such as Anthropic’s Claude Opus 4.6 at 68.8% and OpenAI’s GPT-5.2 at 52.9%. ([the-decoder.com](https://the-decoder.com/google-releases-gemini-3-1-pro-with-improved-reasoning-capabilities/?utm_source=openai))

In the realm of scientific knowledge, Gemini 3.1 Pro scored 94.3% on the GPQA Diamond benchmark, outpacing Claude Opus 4.6’s 91.3% and GPT-5.2’s 92.4%. These results highlight the model’s advanced understanding and processing of complex scientific information. ([officechai.com](https://officechai.com/ai/gemini-3-1-pro-benchmarks/?utm_source=openai))

Advancements in Coding and Agentic Tasks

The model also excels in coding tasks, achieving an 80.6% score on the SWE-Bench Verified benchmark, which evaluates agentic coding capabilities. This performance surpasses Claude Opus 4.6’s 72.6% and GPT-5.3-Codex’s 76.2%. Additionally, Gemini 3.1 Pro leads in agentic tasks, scoring 33.5% on the APEX-Agents benchmark, significantly ahead of GPT-5.2’s 23.0% and Claude Opus 4.6’s 29.8%. ([nxcode.io](https://www.nxcode.io/en/resources/news/gemini-3-1-pro-complete-guide-benchmarks-pricing-api-2026?utm_source=openai))

Cost-Effective Performance

Beyond its technical prowess, Gemini 3.1 Pro offers a cost-effective solution for AI applications. It has claimed the top position on the Artificial Analysis Intelligence Index v4.0 with a score of 57 points, outperforming Claude Opus 4.6 by four points and Claude Sonnet 4.6 by six points. Remarkably, it achieves this at less than half the cost of its nearest competitors, making advanced AI more accessible to a broader range of users. ([officechai.com](https://officechai.com/ai/google-gemini-3-1-pro-takes-top-spot-in-artificial-analysis-intelligence-index-at-price-half-that-of-opus-4-6-gpt-5-2/?utm_source=openai))

Industry Recognition and Availability

The AI community has taken note of Gemini 3.1 Pro’s capabilities. Brendan Foody, CEO of AI startup Mercor, highlighted the model’s performance, stating, Gemini 3.1 Pro is now at the top of the APEX-Agents leaderboard, emphasizing the rapid advancements in AI agents’ ability to perform real-world tasks.

Currently available as a preview, Gemini 3.1 Pro can be accessed through the Gemini API, Google AI Studio, Vertex AI, the Gemini app, and NotebookLM. This wide availability ensures that developers and organizations can integrate the model into various applications, fostering innovation across multiple sectors. ([the-decoder.com](https://the-decoder.com/google-releases-gemini-3-1-pro-with-improved-reasoning-capabilities/?utm_source=openai))

Conclusion

The release of Gemini 3.1 Pro represents a significant milestone in AI development, showcasing Google’s dedication to advancing the field. With its superior performance in reasoning, coding, and scientific knowledge, coupled with cost-effective deployment, Gemini 3.1 Pro sets a new benchmark for AI models, paving the way for future innovations and applications.