Meta's Llama 3 Cracks Top 5 of AI Leaderboard, Only Non-Proprietary Model – Virtualization Review

Spread the love

Advanced Search
News
Meta’s brand-new Llama 3 large language model (LLM) debuted among the top 5 on an AI leaderboard, being the only non-proprietary model.
Meta yesterday (April 18) announced the new open-source model family while describing it as “the most capable openly available LLM to date.”
That openness is a key factor, as the generative AI space has been dominated by proprietary models that are outperforming open-source models, as detailed in a recent report (“AI Report: Closed Models Outperform Open Models, at Staggering Cost“).
In a nutshell, proprietary models usually outperform open-source models typically generated by researchers or government organizations because of the huge and expensive foundation model training required to achieve state-of-the-art performance. Industry players releasing open models is more rare.
Cloud giants like Google and Microsoft have been dominating the AI space in terms of foundation models produced, but Meta has inserted itself among those major players, sandwiched between those two companies.
While Google has led in generating foundation models in 2023, the cost of training its top-level Gemini Ultra model was estimated to near $200 million.
So it’s remarkable that Llama 3 debuted tied at No. 5 in the LMSYS Chatbot Arena Leaderboard.
Chatbot Arena is an open-source research project developed by members from Large Model Systems Organization (LMSYS Org) and UC Berkeley SkyLab, the site says, describing its mission as “to build an open platform to evaluate LLMs by human preference in the real-world.” The crowdsourced open platform for LLM evaluations has collected more than 500,000 human pairwise comparisons to rank LLMs and display the model ratings.
As far as that “most capable” claim, Meta provided this benchmark comparison:
Meta also provided a comparison of Llama 3 with other models in a new human evaluation set, providing real-world rankings, somewhat like the Chatbot Arena:
The company also claimed its pretrained model establishes a new state-of-the-art for LLM models at top-level scales.
No information was revealed regarding training costs for Llama 3. When we asked Meta AI, it said: “Llama 2, a large language model developed by Meta, was trained on a cluster of 512 NVIDIA A100 GPUs for 21 days, with an estimated cost of around $1.8 million,” logically inferring “it’s likely that training Llama 3 would have cost significantly more. For context, Llama 3 has 70 billion parameters, compared to Llama 2’s 35 billion parameters.”
Key takeaways from the Meta announcement as presented by the company include:
Users can also try out the 70 billion-parameter and 8-billion parameter Llama 3 models at AI chip specialist Nvidia’s site.
About the Author
David Ramel is an editor and writer at Converge 360.

Subscribe on YouTube
More Tech Library
More Webcasts
Problems? Questions? Feedback? E-mail us.

source

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top