Anthropic's commanding lead in trader sentiment stems from its Claude Opus 4.6 large language model delivering consistent top-tier results on advanced mathematical reasoning benchmarks such as AIME and HMMT, where it maintains near-perfect accuracy without relying on specialized reasoning modes. This edge reflects Anthropic's focused scaling of chain-of-thought capabilities and long-context handling, which have outpaced competitors in recent independent evaluations. Google and OpenAI trail closely behind due to strong showings from Gemini 3.1 Pro and GPT-5.4 variants on GPQA Diamond and AIME tasks, yet traders view these as slightly less reliable for pure math workloads. With the May resolution window approaching, any fresh benchmark releases or model updates from these labs could still shift the narrow gap between second and third place.
Експериментальне резюме, згенероване ШІ з посиланням на дані Polymarket. Це не торгова порада і не впливає на вирішення цього ринку. · ОновленоWhich company has the best Math AI model end of May?
Anthropic 71%
Google 15%
OpenAI 15%
xAI <1%
$121,102 Обс.
$121,102 Обс.

Anthropic
71%

15%

OpenAI
15%

xAI
1%

ByteDance
1%

Z.ai
1%

DeepSeek
1%

Meta
1%

Baidu
<1%

Alibaba
<1%

Moonshot
<1%

Amazon
<1%

Mistral
<1%

Meituan
<1%

Microsoft
<1%
Anthropic 71%
Google 15%
OpenAI 15%
xAI <1%
$121,102 Обс.
$121,102 Обс.

Anthropic
71%

15%

OpenAI
15%

xAI
1%

ByteDance
1%

Z.ai
1%

DeepSeek
1%

Meta
1%

Baidu
<1%

Alibaba
<1%

Moonshot
<1%

Amazon
<1%

Mistral
<1%

Meituan
<1%

Microsoft
<1%
Results from the "Rank" column under the "Text Arena | Math" Leaderboard tab at https://arena.ai/leaderboard/text/math-no-style-control with style control off will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie still remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies first place under this ranking.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Ринок відкрито: Apr 27, 2026, 5:49 PM ET
Resolver
0x69c47De9D...Results from the "Rank" column under the "Text Arena | Math" Leaderboard tab at https://arena.ai/leaderboard/text/math-no-style-control with style control off will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie still remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies first place under this ranking.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and will resolve based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Resolver
0x69c47De9D...Anthropic's commanding lead in trader sentiment stems from its Claude Opus 4.6 large language model delivering consistent top-tier results on advanced mathematical reasoning benchmarks such as AIME and HMMT, where it maintains near-perfect accuracy without relying on specialized reasoning modes. This edge reflects Anthropic's focused scaling of chain-of-thought capabilities and long-context handling, which have outpaced competitors in recent independent evaluations. Google and OpenAI trail closely behind due to strong showings from Gemini 3.1 Pro and GPT-5.4 variants on GPQA Diamond and AIME tasks, yet traders view these as slightly less reliable for pure math workloads. With the May resolution window approaching, any fresh benchmark releases or model updates from these labs could still shift the narrow gap between second and third place.
Експериментальне резюме, згенероване ШІ з посиланням на дані Polymarket. Це не торгова порада і не впливає на вирішення цього ринку. · Оновлено
Обережно з зовнішніми посиланнями.
Обережно з зовнішніми посиланнями.
Часті запитання