Trader consensus on Anthropic holding an 88.5% implied probability for the best large language model by end of May reflects recent benchmark leadership from its Claude Opus 4.7 and Sonnet 4.6 releases. These models have posted top scores on SWE-Bench Verified for coding and agentic workflows, along with strong human-preference results on Arena leaderboards, outpacing competitors in practical autonomy and instruction following. Google's Gemini 3.1 Pro maintains an 11.5% share through competitive GPQA Diamond reasoning marks, while OpenAI's GPT-5.5 series sits at just 0.5% amid mixed results on extended-context tasks. No major new model launches have shifted the landscape in the past week, though ongoing benchmark updates and any surprise capability demonstrations could still alter market-implied odds before resolution.
Riepilogo sperimentale generato dall'AI con riferimento ai dati di Polymarket. Questo non è un consiglio di trading e non ha alcun ruolo nella risoluzione di questo mercato. · AggiornatoQuale azienda ha il miglior modello di intelligenza artificiale alla fine di maggio?
Anthropic 90%
Google 11%
OpenAI <1%
ByteDance <1%
$8,837,822 Vol.
$8,837,822 Vol.

Anthropic
90%

11%

OpenAI
1%

ByteDance
<1%

Z.ai
<1%

Meta
<1%

xAI
<1%

Alibaba
<1%

Moonshot
<1%

DeepSeek
<1%

Baidu
<1%

Amazon
<1%

Mistral
<1%

Meituan
<1%

Microsoft
<1%
Anthropic 90%
Google 11%
OpenAI <1%
ByteDance <1%
$8,837,822 Vol.
$8,837,822 Vol.

Anthropic
90%

11%

OpenAI
1%

ByteDance
<1%

Z.ai
<1%

Meta
<1%

xAI
<1%

Alibaba
<1%

Moonshot
<1%

DeepSeek
<1%

Baidu
<1%

Amazon
<1%

Mistral
<1%

Meituan
<1%

Microsoft
<1%
Results from the "Rank" column under the "Text Arena | Overall" Leaderboard tab at https://lmarena.ai/leaderboard/text with style control off will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies first place under this ranking system.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Mercato aperto: Apr 14, 2026, 5:17 PM ET
Resolver
0x69c47De9D...Results from the "Rank" column under the "Text Arena | Overall" Leaderboard tab at https://lmarena.ai/leaderboard/text with style control off will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies first place under this ranking system.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Resolver
0x69c47De9D...Trader consensus on Anthropic holding an 88.5% implied probability for the best large language model by end of May reflects recent benchmark leadership from its Claude Opus 4.7 and Sonnet 4.6 releases. These models have posted top scores on SWE-Bench Verified for coding and agentic workflows, along with strong human-preference results on Arena leaderboards, outpacing competitors in practical autonomy and instruction following. Google's Gemini 3.1 Pro maintains an 11.5% share through competitive GPQA Diamond reasoning marks, while OpenAI's GPT-5.5 series sits at just 0.5% amid mixed results on extended-context tasks. No major new model launches have shifted the landscape in the past week, though ongoing benchmark updates and any surprise capability demonstrations could still alter market-implied odds before resolution.
Riepilogo sperimentale generato dall'AI con riferimento ai dati di Polymarket. Questo non è un consiglio di trading e non ha alcun ruolo nella risoluzione di questo mercato. · Aggiornato
Fai attenzione ai link esterni.
Fai attenzione ai link esterni.
Domande frequenti