OpenAI's GPT-5.5 Pro recently claimed the FrontierMath lead at 52.4% as of May 13, 2026, outpacing Anthropic's Claude Opus 4.7 Adaptive at 43.8% on this benchmark of 350 expert-vetted math problems spanning undergraduate to research-level challenges, evaluated with Python tool access. Claude's score reflects solid progress from prior versions but trails amid OpenAI's rapid math reasoning advances, highlighting Anthropic's relative emphasis on agentic coding strengths like 77.8% on SWE-Bench Pro over pure mathematical scaling. With six weeks until June 30 resolution, traders eye potential Claude 5 previews or Opus updates, bolstered by Anthropic's $1.8 billion Akamai compute deal, though Epoch AI's recent flagging of benchmark errors adds evaluation uncertainty.
Экспериментальная сводка, созданная ИИ на основе данных Polymarket. Это не является торговой рекомендацией и не влияет на то, как разрешается этот рынок. · ОбновленоАнтропный балл Клода по FrontierMath Benchmark к 30 июня?
Антропный балл Клода по FrontierMath Benchmark к 30 июня?
$61,907 Объем
50%+
55%
$61,907 Объем
50%+
55%
This market will resolve according to the Epoch AI’s Frontier Math benchmarking leaderboard (https://epoch.ai/frontiermath) for Tier 1-3. Studies which are not included in the leaderboard (e.g. https://x.com/EpochAIResearch/status/1945905796904005720) will not be considered.
The primary resolution source will be information from EpochAI; however, a consensus of credible reporting may also be used.
Открытие рынка: Jan 30, 2026, 12:00 AM ET
Resolver
0x65070BE91...This market will resolve according to the Epoch AI’s Frontier Math benchmarking leaderboard (https://epoch.ai/frontiermath) for Tier 1-3. Studies which are not included in the leaderboard (e.g. https://x.com/EpochAIResearch/status/1945905796904005720) will not be considered.
The primary resolution source will be information from EpochAI; however, a consensus of credible reporting may also be used.
Resolver
0x65070BE91...OpenAI's GPT-5.5 Pro recently claimed the FrontierMath lead at 52.4% as of May 13, 2026, outpacing Anthropic's Claude Opus 4.7 Adaptive at 43.8% on this benchmark of 350 expert-vetted math problems spanning undergraduate to research-level challenges, evaluated with Python tool access. Claude's score reflects solid progress from prior versions but trails amid OpenAI's rapid math reasoning advances, highlighting Anthropic's relative emphasis on agentic coding strengths like 77.8% on SWE-Bench Pro over pure mathematical scaling. With six weeks until June 30 resolution, traders eye potential Claude 5 previews or Opus updates, bolstered by Anthropic's $1.8 billion Akamai compute deal, though Epoch AI's recent flagging of benchmark errors adds evaluation uncertainty.
Экспериментальная сводка, созданная ИИ на основе данных Polymarket. Это не является торговой рекомендацией и не влияет на то, как разрешается этот рынок. · Обновлено
Не доверяй внешним ссылкам.
Не доверяй внешним ссылкам.
Часто задаваемые вопросы