Navigation
Recherche
|
Study Accuses LM Arena of Helping Top AI Labs Game Its Benchmark
jeudi 1 mai 2025, 15:00 , par Slashdot
![]() According to the authors, LM Arena allowed some industry-leading AI companies like Meta, OpenAI, Google, and Amazon to privately test several variants of AI models, then not publish the scores of the lowest performers. This made it easier for these companies to achieve a top spot on the platform's leaderboard, though the opportunity was not afforded to every firm, the authors say. 'Only a handful of [companies] were told that this private testing was available, and the amount of private testing that some [companies] received is just so much more than others,' said Cohere's VP of AI research and co-author of the study, Sara Hooker, in an interview with TechCrunch. 'This is gamification.' Further reading: Meta Got Caught Gaming AI Benchmarks. Read more of this story at Slashdot.
https://slashdot.org/story/25/05/01/0525208/study-accuses-lm-arena-of-helping-top-ai-labs-game-its-b...
Voir aussi |
56 sources (32 en français)
Date Actuelle
ven. 2 mai - 14:05 CEST
|