Anthropic’s Claude Opus 4.6 leads trader consensus at 62.5% implied probability because its recent reasoning optimizations deliver top scores on rigorous math benchmarks such as MATH-500 and AIME, where extended chain-of-thought processing outperforms prior generations. OpenAI’s GPT-5 family sits at 20.5% on the strength of consistent high-90s results across competition-style problems, while Google’s Gemini 3.1 Pro holds 17.5% after closing much of the gap with its multi-tier thinking mode. Low single-digit odds for xAI, ByteDance, and others reflect limited public math-specific gains this month. With resolution at month-end tied to the latest verifiable leaderboards, any new benchmark release or capability demonstration could still shift positioning before the final tally.
Riepilogo sperimentale generato dall'AI con riferimento ai dati di Polymarket. Questo non è un consiglio di trading e non ha alcun ruolo nella risoluzione di questo mercato. · AggiornatoWhich company has the best Math AI model end of May?
Anthropic 63%
OpenAI 21%
Google 18%
xAI <1%
$114,284 Vol.
$114,284 Vol.

Anthropic
63%

OpenAI
21%

18%

xAI
1%

ByteDance
1%

Z.ai
1%

DeepSeek
1%

Meta
1%

Baidu
<1%

Alibaba
<1%

Moonshot
<1%

Amazon
<1%

Mistral
<1%

Meituan
<1%

Microsoft
<1%
Anthropic 63%
OpenAI 21%
Google 18%
xAI <1%
$114,284 Vol.
$114,284 Vol.

Anthropic
63%

OpenAI
21%

18%

xAI
1%

ByteDance
1%

Z.ai
1%

DeepSeek
1%

Meta
1%

Baidu
<1%

Alibaba
<1%

Moonshot
<1%

Amazon
<1%

Mistral
<1%

Meituan
<1%

Microsoft
<1%
Results from the "Rank" column under the "Text Arena | Math" Leaderboard tab at https://arena.ai/leaderboard/text/math-no-style-control with style control off will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie still remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies first place under this ranking.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Mercato aperto: Apr 27, 2026, 5:49 PM ET
Resolver
0x69c47De9D...Results from the "Rank" column under the "Text Arena | Math" Leaderboard tab at https://arena.ai/leaderboard/text/math-no-style-control with style control off will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie still remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies first place under this ranking.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Resolver
0x69c47De9D...Anthropic’s Claude Opus 4.6 leads trader consensus at 62.5% implied probability because its recent reasoning optimizations deliver top scores on rigorous math benchmarks such as MATH-500 and AIME, where extended chain-of-thought processing outperforms prior generations. OpenAI’s GPT-5 family sits at 20.5% on the strength of consistent high-90s results across competition-style problems, while Google’s Gemini 3.1 Pro holds 17.5% after closing much of the gap with its multi-tier thinking mode. Low single-digit odds for xAI, ByteDance, and others reflect limited public math-specific gains this month. With resolution at month-end tied to the latest verifiable leaderboards, any new benchmark release or capability demonstration could still shift positioning before the final tally.
Riepilogo sperimentale generato dall'AI con riferimento ai dati di Polymarket. Questo non è un consiglio di trading e non ha alcun ruolo nella risoluzione di questo mercato. · Aggiornato
Fai attenzione ai link esterni.
Fai attenzione ai link esterni.
Domande frequenti