Roche asks:
Which organization's large language model (LLM) will be ranked first as of 29 May 2026, according to Arena's "Text Arena" leaderboard?
Closing May 29, 2026 07:01AM UTC
Arena is an open-source platform for crowdsourced AI benchmarking, created by researchers from UC Berkeley SkyLab (SkyLab). For more information on how the leaderboard is constructed, see Arena - Blog. The question will be suspended on 28 May 2026 and the outcome determined using the ranks as reported by Arena at approximately 5:00p.m. ET on 29 May 2026 (Arena - Text Arena Leaderboard, see "Rank"). As of 5 February 2026, Google was ranked first, with its "gemini-3-pro" scoring 1487, followed by xAI's "grok-4.1-thinking" scoring 1475. In the event of a tie for first place by LLMs of different organizations, the LLM with the higher "Score" will be considered first, followed by the "Votes" total. If the named source changes the way it presents the data, further instructions will be provided. Models marked "Preliminary" next to their score will not count.
Confused? Check our FAQ or ask us for help. To learn more about Good Judgment and Superforecasting, click here.
To learn more about how you can become a Superforecaster, see here. For other posts from our Insights blog, click here.