Anthropic’s Claude Opus 4.7 and Sonnet 4.6 continue to drive the 68% market-implied odds through demonstrated strengths in instruction-following, nuanced writing style, and agentic coding tasks that align closely with the “Style Control On” evaluation criteria. Recent benchmark updates show these models leading SWE-bench Verified and LM Arena text categories while maintaining a 1-million-token context window, giving traders confidence in sustained leadership through June. Google’s Gemini 3.1 Pro Preview, at 23%, has posted top scores on GPQA Diamond and ARC-AGI-2 but trails in style-specific metrics. OpenAI’s April GPT-5.5 release and xAI’s Grok 4.3 update have not yet shifted the consensus, as traders weigh competitive positioning and upcoming developer events that could still influence final rankings.
Експериментальне резюме, згенероване ШІ з посиланням на дані Polymarket. Це не торгова порада і не впливає на вирішення цього ринку. · ОновленоЯка компанія має найкращу модель ШІ наприкінці червня? (Керування стилем увімкнено)
Anthropic 68%
Google 23%
OpenAI 6%
xAI 2.6%
$1,457,332 Обс.
$1,457,332 Обс.

Anthropic
68%

23%

OpenAI
6%

xAI
3%

DeepSeek
1%

Alibaba
1%

Meta
1%

Z.ai
1%

Microsoft
1%

Moonshot
1%

Meituan
<1%

Mistral
<1%

ByteDance
<1%

Baidu
<1%

Amazon
<1%
Anthropic 68%
Google 23%
OpenAI 6%
xAI 2.6%
$1,457,332 Обс.
$1,457,332 Обс.

Anthropic
68%

23%

OpenAI
6%

xAI
3%

DeepSeek
1%

Alibaba
1%

Meta
1%

Z.ai
1%

Microsoft
1%

Moonshot
1%

Meituan
<1%

Mistral
<1%

ByteDance
<1%

Baidu
<1%

Amazon
<1%
Results from the "Rank" column under the "Text Arena | Overall" Leaderboard tab at https://lmarena.ai/leaderboard/text with style control on will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies first place under this ranking system.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and resolves based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Ринок відкрито: Oct 10, 2025, 5:27 PM ET
Resolver
0x2F5e3684c...Results from the "Rank" column under the "Text Arena | Overall" Leaderboard tab at https://lmarena.ai/leaderboard/text with style control on will be used to resolve this market.
Models will be ordered primarily by their leaderboard rank at the market’s check time. If two or more models are tied on rank, they will be ordered by their Arena score, including any underlying, unrounded, granular values reflected in the data below the leaderboard. If a tie remains, alphabetical order of company names as listed in this market group will be used as a final tiebreaker (e.g., if the two models are tied by exact arena score, “Google” would be ranked ahead of “xAI”). This market will resolve based on the company that occupies first place under this ranking system.
The resolution source for this market is the Chatbot Arena LLM Leaderboard found at https://lmarena.ai/. If this resolution source is unavailable at check time, this market will remain open until the leaderboard comes back online and resolves based on the first check after it becomes available. If it becomes permanently unavailable, this market will resolve based on another resolution source.
Resolver
0x2F5e3684c...Anthropic’s Claude Opus 4.7 and Sonnet 4.6 continue to drive the 68% market-implied odds through demonstrated strengths in instruction-following, nuanced writing style, and agentic coding tasks that align closely with the “Style Control On” evaluation criteria. Recent benchmark updates show these models leading SWE-bench Verified and LM Arena text categories while maintaining a 1-million-token context window, giving traders confidence in sustained leadership through June. Google’s Gemini 3.1 Pro Preview, at 23%, has posted top scores on GPQA Diamond and ARC-AGI-2 but trails in style-specific metrics. OpenAI’s April GPT-5.5 release and xAI’s Grok 4.3 update have not yet shifted the consensus, as traders weigh competitive positioning and upcoming developer events that could still influence final rankings.
Експериментальне резюме, згенероване ШІ з посиланням на дані Polymarket. Це не торгова порада і не впливає на вирішення цього ринку. · Оновлено
Обережно з зовнішніми посиланнями.
Обережно з зовнішніми посиланнями.
Часті запитання