Google DeepMind’s Gemini 3.1 Pro series and its new agentic “Co-Mathematician” workbench form the core driver behind trader sentiment on the FrontierMath market. The benchmark, built from hundreds of unpublished research-level math problems, currently shows OpenAI’s GPT-5.4 at the top with 47.6 percent accuracy, while baseline Gemini 3.1 Pro sits near 19 percent on the hardest Tier 4 subset. Recent agentic enhancements have lifted Gemini performance to 48 percent on Tier 4 in internal tests, narrowing the gap that earlier Gemini 3 “Deep Think” runs had closed to roughly 37–40 percent overall. With June 30 only weeks away, traders are watching for any official Epoch AI re-evaluation or a rapid Gemini update that could push the model into the low-to-mid 40s before resolution.
Resumo experimental gerado por IA com dados do Polymarket. Isto não é aconselhamento de trading e não tem qualquer papel na resolução deste mercado. · AtualizadoPontuação do Google Gemini no FrontierMath Benchmark até 30 de junho?
$136,324 Vol.
40%+
86%
45%+
63%
50%+
63%
60%+
54%
$136,324 Vol.
40%+
86%
45%+
63%
50%+
63%
60%+
54%
This market will resolve according to the Epoch AI’s Frontier Math benchmarking leaderboard (https://epoch.ai/frontiermath) for Tier 1-3. Studies which are not included in the leaderboard (e.g. https://x.com/EpochAIResearch/status/1945905796904005720) will not be considered.
The primary resolution source will be information from EpochAI; however, a consensus of credible reporting may also be used.
Mercado Aberto: Feb 6, 2026, 6:03 PM ET
Resolver
0x65070BE91...This market will resolve according to the Epoch AI’s Frontier Math benchmarking leaderboard (https://epoch.ai/frontiermath) for Tier 1-3. Studies which are not included in the leaderboard (e.g. https://x.com/EpochAIResearch/status/1945905796904005720) will not be considered.
The primary resolution source will be information from EpochAI; however, a consensus of credible reporting may also be used.
Resolver
0x65070BE91...Google DeepMind’s Gemini 3.1 Pro series and its new agentic “Co-Mathematician” workbench form the core driver behind trader sentiment on the FrontierMath market. The benchmark, built from hundreds of unpublished research-level math problems, currently shows OpenAI’s GPT-5.4 at the top with 47.6 percent accuracy, while baseline Gemini 3.1 Pro sits near 19 percent on the hardest Tier 4 subset. Recent agentic enhancements have lifted Gemini performance to 48 percent on Tier 4 in internal tests, narrowing the gap that earlier Gemini 3 “Deep Think” runs had closed to roughly 37–40 percent overall. With June 30 only weeks away, traders are watching for any official Epoch AI re-evaluation or a rapid Gemini update that could push the model into the low-to-mid 40s before resolution.
Resumo experimental gerado por IA com dados do Polymarket. Isto não é aconselhamento de trading e não tem qualquer papel na resolução deste mercado. · Atualizado
Cuidado com os links externos.
Cuidado com os links externos.
Frequently Asked Questions