Meta releases open-source big model Llama 3.1 with strong support from Nvidia
网事大话每
发表于 2024-7-24 13:05:30
1249
0
0
Science and Technology Innovation Board Daily, July 24th (Reporter Zhang Yangyang) Zuckerberg will continue to open source big models to the end.
Early this morning, Meta officially released the new generation of open-source big model Llama 3.1 series, which includes three versions: 8B, 70B, and 405B, with a maximum context increase of 128k.
Meta founder Mark Zuckerberg also posted on the official website to strongly endorse his own model. He said that most leading technology companies and scientific research today are built on open source software, which is the direction for AI to move forward, and Meta is moving towards becoming the industry standard for open source AI.
It should be emphasized that in the technology industry, the dispute over open source and closed source has a long history. Critics argue that open source conceals a lack of technological originality and only makes simple adjustments to the open source model, rather than substantive innovation. Robin Lee, the founder of Baidu, even said that the open source model has value in academic research, teaching and other specific scenarios, but it is not applicable to most application scenarios. Supporters believe that customized improvements based on mature open source architectures are the norm of technological development, which can drive rapid innovation and progress in technology.
In the field of big models, there is often a comparison of the advantages and disadvantages between open source and closed source big models. So far, open-source models have mostly lagged behind closed models in terms of functionality and performance. But with the release of Llama 3.1, there may be a new round of intense competition between open source and closed source big models.
According to benchmark data provided by Meta, Llama 3.1 has 405 billion parameters, making it one of the largest large-scale language models in recent years. This model is trained on 15 trillion tokens and over 16000 H100 GPUs, making it the first Llama model in Meta's history to be trained on this scale. Meta states that in terms of advanced features such as common sense, manipulability, mathematics, tool usage, and multilingual translation, Llama 3.1 is sufficient to benchmark top closed source big models such as GPT-4o and Claude 3.5Sonnet.
Llama 3.1 is now available for download on the Meta official website and Hugging Face. The latest data shows that the total download volume of all Llama versions has exceeded 300 million times.
At the same time on the same day, Nvidia also launched a combination training service, providing strong assists for Llama 3.1.
The reporter from the Science and Technology Innovation Board Daily learned from Nvidia that Nvidia has officially launched new NVIDIA AI Foundry services and NVIDIA NIM inference microservices. NVIDIA AI Foundry is driven by the NVIDIA DGX Cloud AI platform, which is jointly designed by NVIDIA and public cloud and can provide enterprises with a large amount of computing power resources.
NVIDIA AI Foundry and NVIDIA NIM are used together with the Llama 3.1 series open source models, allowing enterprises to create custom "super models" for their specific industry use cases. Enterprises can also use their own data and synthetic data generated by Llama 3.1 405B and NVIDIA Nemotron Reward models to train these super models.
Nvidia founder and CEO Huang Renxun stated that Meta's Llama 3.1 open-source model marks a critical moment for global enterprises to adopt generative AI. Llama 3.1 will ignite a wave of enterprises and industries creating advanced generative AI applications. NVIDIA AI Foundry has integrated Llama 3.1 throughout the entire process and is able to assist enterprises in building and deploying custom Llama hypermodels.
CandyLake.com 系信息发布平台,仅提供信息存储空间服务。
声明:该文观点仅代表作者本人,本文不代表CandyLake.com立场,且不构成建议,请谨慎对待。
声明:该文观点仅代表作者本人,本文不代表CandyLake.com立场,且不构成建议,请谨慎对待。
猜你喜欢
- Strong demand for AI drives Broadcom's profits beyond expectations, causing its stock price to soar 15% after hours
- Can Broadcom's customized AI chip challenge Nvidia with a market value exceeding trillions of dollars?
- Nvidia's US stock fell more than 2% before trading
- Nvidia's stock price fell 2.1% in pre-market trading and is expected to decline for four consecutive trading days
- Who will dominate the new landscape of AI chips between Broadcom and Nvidia?
- Who is the biggest buyer of Nvidia AI chips? This tech giant is dominating the rankings ahead of its peers
- Nvidia's US stock rose over 2% in pre-market trading
- Research institution: Microsoft will purchase far more Nvidia AI chips than its competitors in 2024
- Nvidia's stock price rose 2.5% in pre-market trading and is expected to end its four consecutive declines
- Strong entry! JD is expected to obtain a consumer finance license
-
生成式人工知能(AI)が巻き起こす技術の波の中で、電力会社は意外にも資本市場の寵児になった。 今年のスタンダード500割株の上昇幅ランキングでは、Vistraなどの従来の電力会社が注目を集め、株価が2倍になってリ ...
- xifangczy
- 3 天前
- 支持
- 反对
- 回复
- 收藏
-
隔夜株式市場 世界の主要指数は金曜日に多くが下落し、最新のインフレデータが減速の兆しを示したおかげで、米株3大指数は大幅に回復し、いずれも1%超上昇した。 金曜日に発表されたデータによると、米国の11月のPC ...
- SNT
- 前天 12:48
- 支持
- 反对
- 回复
- 收藏
-
長年にわたって、昔の消金大手の捷信消金の再編がようやく地に着いた。 天津銀行の発表によると、同行は京東傘下の2社、対外貿易信託などと捷信消金再編に参加する。再編が完了すると、京東の持ち株比率は65%に達し ...
- SNT
- 前天 12:09
- 支持
- 反对
- 回复
- 收藏
-
グーグルは現地時間12月19日、新しい「推理」モデルとしてGemini 2.0 Flash Thinkingを発売すると発表した。紹介によると、このモデルはまだ実験段階であり、訓練を経た後、モデルが反応を起こした時に経験した「思 ...
- 地下水
- 3 天前
- 支持
- 反对
- 回复
- 收藏