Google’s Gemini family continues to demonstrate competitive strength in advanced mathematical reasoning on the FrontierMath benchmark, with Gemini 3 Pro posting 37.6 percent accuracy and its “Deep Think” variant exceeding 40 percent in late 2025 evaluations. The February 2026 release of Gemini 3.1 Pro achieved comparable results on the Tier 1–3 problems while becoming the first model to solve a previously unsolved Tier 4 problem, underscoring incremental gains in handling unpublished, research-level mathematics from number theory to algebraic geometry. OpenAI’s GPT-5.4 currently leads the leaderboard at 47.6 percent, creating a tight competitive gap that traders are watching closely. With June 30 approaching, any unreleased Gemini update, internal scaling improvements, or refined inference techniques could shift the final score before the market resolves.
Polymarketデータを参照したAI生成の実験的な要約。これは取引アドバイスではなく、このマーケットの解決方法には一切関係ありません。 · 更新日$136,324 Vol.
40%以上
86%
45%以上
66%
50%以上
62%
60%以上
54%
$136,324 Vol.
40%以上
86%
45%以上
66%
50%以上
62%
60%以上
54%
This market will resolve according to the Epoch AI’s Frontier Math benchmarking leaderboard (https://epoch.ai/frontiermath) for Tier 1-3. Studies which are not included in the leaderboard (e.g. https://x.com/EpochAIResearch/status/1945905796904005720) will not be considered.
The primary resolution source will be information from EpochAI; however, a consensus of credible reporting may also be used.
マーケット開始日: Feb 6, 2026, 6:03 PM ET
Resolver
0x65070BE91...This market will resolve according to the Epoch AI’s Frontier Math benchmarking leaderboard (https://epoch.ai/frontiermath) for Tier 1-3. Studies which are not included in the leaderboard (e.g. https://x.com/EpochAIResearch/status/1945905796904005720) will not be considered.
The primary resolution source will be information from EpochAI; however, a consensus of credible reporting may also be used.
Resolver
0x65070BE91...Google’s Gemini family continues to demonstrate competitive strength in advanced mathematical reasoning on the FrontierMath benchmark, with Gemini 3 Pro posting 37.6 percent accuracy and its “Deep Think” variant exceeding 40 percent in late 2025 evaluations. The February 2026 release of Gemini 3.1 Pro achieved comparable results on the Tier 1–3 problems while becoming the first model to solve a previously unsolved Tier 4 problem, underscoring incremental gains in handling unpublished, research-level mathematics from number theory to algebraic geometry. OpenAI’s GPT-5.4 currently leads the leaderboard at 47.6 percent, creating a tight competitive gap that traders are watching closely. With June 30 approaching, any unreleased Gemini update, internal scaling improvements, or refined inference techniques could shift the final score before the market resolves.
Polymarketデータを参照したAI生成の実験的な要約。これは取引アドバイスではなく、このマーケットの解決方法には一切関係ありません。 · 更新日
外部リンクに注意してください。
外部リンクに注意してください。
よくある質問