Recent model releases and benchmark results have solidified trader consensus around Anthropic as the third-strongest large language model provider by end of May. OpenAI’s GPT-5.5 Instant and Pro variants, rolled out in early May, deliver strong all-round performance and reduced hallucinations, while Google’s Gemini 3.1 Pro maintains clear leads on reasoning metrics such as GPQA Diamond and ARC-AGI-2. Anthropic’s Claude Opus 4.7 and Sonnet 4.6 updates, with 1-million-token context and top-tier coding scores on SWE-bench, keep the company competitive yet consistently behind these two on aggregated evaluations. This positioning, reinforced by mid-May developer feedback and leaderboard data, leaves little room for other labs to claim third place before month-end.
基于Polymarket数据的AI实验性摘要。这不是交易建议,也不影响该市场的结算方式。 · 更新于Anthropic 70%
Google 31%
OpenAI 1.3%
xAI <1%
$91,800 交易量
$91,800 交易量

Anthropic
70%

31%

OpenAI
1%

xAI
1%

Baidu
<1%

Meta
<1%

Z.ai
<1%

ByteDance
<1%

Alibaba
<1%

Moonshot
<1%

Meituan
<1%

DeepSeek
<1%

Microsoft
<1%

Amazon
<1%

Mistral
<1%
Anthropic 70%
Google 31%
OpenAI 1.3%
xAI <1%
$91,800 交易量
$91,800 交易量

Anthropic
70%

31%

OpenAI
1%

xAI
1%

Baidu
<1%

Meta
<1%

Z.ai
<1%

ByteDance
<1%

Alibaba
<1%

Moonshot
<1%

Meituan
<1%

DeepSeek
<1%

Microsoft
<1%

Amazon
<1%

Mistral
<1%
Results from the "Rank" column under the "Text Arena | Overall" Leaderboard tab at https://lmarena.ai/leaderboard/text with style control off will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies third place under this ranking system.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
市场开放时间: Apr 14, 2026, 5:18 PM ET
Resolver
0x69c47De9D...Results from the "Rank" column under the "Text Arena | Overall" Leaderboard tab at https://lmarena.ai/leaderboard/text with style control off will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies third place under this ranking system.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Resolver
0x69c47De9D...Recent model releases and benchmark results have solidified trader consensus around Anthropic as the third-strongest large language model provider by end of May. OpenAI’s GPT-5.5 Instant and Pro variants, rolled out in early May, deliver strong all-round performance and reduced hallucinations, while Google’s Gemini 3.1 Pro maintains clear leads on reasoning metrics such as GPQA Diamond and ARC-AGI-2. Anthropic’s Claude Opus 4.7 and Sonnet 4.6 updates, with 1-million-token context and top-tier coding scores on SWE-bench, keep the company competitive yet consistently behind these two on aggregated evaluations. This positioning, reinforced by mid-May developer feedback and leaderboard data, leaves little room for other labs to claim third place before month-end.
基于Polymarket数据的AI实验性摘要。这不是交易建议,也不影响该市场的结算方式。 · 更新于
警惕外部链接哦。
警惕外部链接哦。
常见问题