Will general purpose AI models beat average score of human players in Diplomacy by 2028?
Plus
10
Ṁ11712027
60%
chance
1D
1W
1M
ALL
General purpose (not trained for a specific task) language models demonstrated chess playing ability. They are also capable of deception and lie detection. Will language models or visual-language models* beat the average score of human players during a series of 40 games on webDiplomacy.net by 2028? (question modeled after Meta's Cicero result).
[EDIT: Please notice that while "CICERO achieved more than 2x the average score of its opponents" this question requires only achieving the above-average score]
*models or agents trained on different modalities (so e. g. models capable of controlling robotic arm like PaLM-E) would also qualify as long as they weren't trained specifically to play Diplomacy
Get Ṁ1,000 play money
Related questions
Related questions
In 2028, will an AI be able to play randomly selected computer games at human level without getting to practice?
60% chance
Will any AI get a score of at least 45% on Humanity’s Last Exam benchmark before March 11, 2025?
12% chance
Will AI beat top human players at Civ6 (without cheating) by EOY 2026?
32% chance
Will an AI score 1st place on International Math Olympiad (IMO) 2025?
28% chance
Will AI image generating models score >= 90% on Winoground by June 1, 2025?
82% chance
Will an AI model achieve superhuman ELO on Codeforces by the 31 December 2025?
61% chance
Will an AI model outperform 95% of Manifold users on accuracy before 2026?
58% chance
Will an AI be capable of achieving a perfect score on the Putnam exam before 2026?
42% chance
Will any AI model score >80% on Epoch's Frontier Math Benchmark in 2025?
26% chance
Will an AI system beat humans in the GAIA benchmark before the end of 2025?
60% chance