Meta’s benchmarks for its new AI models are a bit misleading

Meta app icon in 3D

Meta app icon in 3D

Meta's recent flagship AI model, Maverick, has sparked interest and suspicion.

While Maverick ranks remarkably high on LM Arena, a platform that uses human feedback to compare AI models, there's a catch: The version tested is not the same one developers can actually use.

Meta 3D logo
expand image

Meta has quietly revealed that the version submitted to LM Arena is an "experimental chat version" designed exclusively for conversational performance. The official Llama website describes this version as "Llama 4 Maverick optimized for conversationality." In other words, it's a customized, souped-up variation designed to do well in the LM Arena evaluation format. The publicly available version? In some scenarios, a more generalized, "vanilla" model is likely to be less flashy and less capable.

This difference causes major worry. Benchmarks are meant to represent a model's strengths and flaws accurately. Customizing a model expressly to perform well in a benchmark contradicts the test's purpose. It's like test-driving a performance-enhanced car only to get the standard model when you purchase it.

Developers that rely on the LM Arena results may be in for a surprise. Researchers have flagged significant behavioral differences between the LM Arena Maverick and the downloaded version, ranging from excessive emoji usage to overly lengthy responses. These changes aren't just cosmetic; they suggest a deeper divergence in how the models trend or tweet.

The issue is transparency. If Meta works to optimize for benchmarks behind the scenes while providing different models to the public, it clouds expectations and trust. While adapting for benchmarks isn't new, it's often frowned upon unless fully revealed, and Meta's tiny print falls short.

Man using new meta threads app on smartphone.
expand image

Accurate benchmarking is essential as AI becomes more integral to industries like crypto and decentralized tech. Developers, investors, and users must understand precisely what they are getting. Maverick may be a powerful model, but Meta's representation of it feels more like a public relations stunt than a true study, and that's a problem.