Recent releases from Anthropic have accelerated progress in the Code Arena benchmark, with Claude Opus 4.7 variants reaching scores above 1570 through improved agentic planning and frontend execution on real web-development tasks. This has widened the gap over OpenAI’s GPT-5.4 series and Google’s Gemini 3.1 Pro, which trail by 80–100 Elo points despite strong SWE-bench results. With only six weeks until June 30, traders focus on whether incremental updates or new model drops from xAI, Meta, or open-weight labs can close that distance before the deadline. Historical benchmark gains suggest rapid iteration remains possible, but the compressed timeline limits upside beyond current leaderboards.
Polymarketデータを参照したAI生成の実験的な要約。これは取引アドバイスではなく、このマーケットの解決方法には一切関係ありません。 · 更新日1550
54%
1560
59%
1570
26%
$7,730 Vol.
1550
54%
1560
59%
1570
26%
Results from the "Score" column under the "Text Arena | Coding" Leaderboard tab at https://arena.ai/leaderboard/text/coding-no-style-control with style control off will be used to resolve this market.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at arena.ai/leaderboard/text. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If permanently unavailable, this market will resolve to "No".
マーケット開始日: Apr 2, 2026, 6:09 PM ET
Resolver
0x65070BE91...Results from the "Score" column under the "Text Arena | Coding" Leaderboard tab at https://arena.ai/leaderboard/text/coding-no-style-control with style control off will be used to resolve this market.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at arena.ai/leaderboard/text. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If permanently unavailable, this market will resolve to "No".
Resolver
0x65070BE91...Recent releases from Anthropic have accelerated progress in the Code Arena benchmark, with Claude Opus 4.7 variants reaching scores above 1570 through improved agentic planning and frontend execution on real web-development tasks. This has widened the gap over OpenAI’s GPT-5.4 series and Google’s Gemini 3.1 Pro, which trail by 80–100 Elo points despite strong SWE-bench results. With only six weeks until June 30, traders focus on whether incremental updates or new model drops from xAI, Meta, or open-weight labs can close that distance before the deadline. Historical benchmark gains suggest rapid iteration remains possible, but the compressed timeline limits upside beyond current leaderboards.
Polymarketデータを参照したAI生成の実験的な要約。これは取引アドバイスではなく、このマーケットの解決方法には一切関係ありません。 · 更新日
外部リンクに注意してください。
外部リンクに注意してください。
よくある質問