Meta releases open-source big model Llama 3.1 with strong support from Nvidia
网事大话每
发表于 2024-7-24 13:05:30
1236
0
0
Science and Technology Innovation Board Daily, July 24th (Reporter Zhang Yangyang) Zuckerberg will continue to open source big models to the end.
Early this morning, Meta officially released the new generation of open-source big model Llama 3.1 series, which includes three versions: 8B, 70B, and 405B, with a maximum context increase of 128k.
Meta founder Mark Zuckerberg also posted on the official website to strongly endorse his own model. He said that most leading technology companies and scientific research today are built on open source software, which is the direction for AI to move forward, and Meta is moving towards becoming the industry standard for open source AI.
It should be emphasized that in the technology industry, the dispute over open source and closed source has a long history. Critics argue that open source conceals a lack of technological originality and only makes simple adjustments to the open source model, rather than substantive innovation. Robin Lee, the founder of Baidu, even said that the open source model has value in academic research, teaching and other specific scenarios, but it is not applicable to most application scenarios. Supporters believe that customized improvements based on mature open source architectures are the norm of technological development, which can drive rapid innovation and progress in technology.
In the field of big models, there is often a comparison of the advantages and disadvantages between open source and closed source big models. So far, open-source models have mostly lagged behind closed models in terms of functionality and performance. But with the release of Llama 3.1, there may be a new round of intense competition between open source and closed source big models.
According to benchmark data provided by Meta, Llama 3.1 has 405 billion parameters, making it one of the largest large-scale language models in recent years. This model is trained on 15 trillion tokens and over 16000 H100 GPUs, making it the first Llama model in Meta's history to be trained on this scale. Meta states that in terms of advanced features such as common sense, manipulability, mathematics, tool usage, and multilingual translation, Llama 3.1 is sufficient to benchmark top closed source big models such as GPT-4o and Claude 3.5Sonnet.
Llama 3.1 is now available for download on the Meta official website and Hugging Face. The latest data shows that the total download volume of all Llama versions has exceeded 300 million times.
At the same time on the same day, Nvidia also launched a combination training service, providing strong assists for Llama 3.1.
The reporter from the Science and Technology Innovation Board Daily learned from Nvidia that Nvidia has officially launched new NVIDIA AI Foundry services and NVIDIA NIM inference microservices. NVIDIA AI Foundry is driven by the NVIDIA DGX Cloud AI platform, which is jointly designed by NVIDIA and public cloud and can provide enterprises with a large amount of computing power resources.
NVIDIA AI Foundry and NVIDIA NIM are used together with the Llama 3.1 series open source models, allowing enterprises to create custom "super models" for their specific industry use cases. Enterprises can also use their own data and synthetic data generated by Llama 3.1 405B and NVIDIA Nemotron Reward models to train these super models.
Nvidia founder and CEO Huang Renxun stated that Meta's Llama 3.1 open-source model marks a critical moment for global enterprises to adopt generative AI. Llama 3.1 will ignite a wave of enterprises and industries creating advanced generative AI applications. NVIDIA AI Foundry has integrated Llama 3.1 throughout the entire process and is able to assist enterprises in building and deploying custom Llama hypermodels.
CandyLake.com is an information publishing platform and only provides information storage space services.
Disclaimer: The views expressed in this article are those of the author only, this article does not represent the position of CandyLake.com, and does not constitute advice, please treat with caution.
Disclaimer: The views expressed in this article are those of the author only, this article does not represent the position of CandyLake.com, and does not constitute advice, please treat with caution.
You may like
- Private equity leader Jinglin's US stock holdings exposed: selling off Nvidia, Microsoft adding positions in Apple, Tencent Music, etc
- AI Weekly | AI search startup Perplexity valuation surges to $9 billion; Nvidia's market value exceeds Apple's
- Is AI unstoppable? NVIDIA's total market value exceeds $3.6 trillion
- Stock price rises to beat Nvidia! What insights does the emergence of AI applications in the US stock market bring to A-shares?
- Nvidia announces partnership with SoftBank to build AI infrastructure in Japan
- Nvidia and SoftBank test run the world's first artificial intelligence and 5G telecommunications network
- The world's first! Nvidia and SoftBank's trial operation of "AI+5G" marks a major breakthrough in the telecommunications industry
- Alibaba Tongyi Qianwen Code Model Qwen2.5-Coder Full Series Officially Open Source
- Citigroup raises Nvidia target price from $150 to $170
- Nvidia's stock price can rise another 26%. Analysts say 'iPhone moment' is coming soon!
-
11월 14일, 세계예선 아시아지역 제3단계 C조 제5라운드, 중국남자축구는 바레인남자축구와 원정경기를 가졌다.축구 국가대표팀은 바레인을 1-0으로 꺾고 예선 2연승을 거두었다. 특히 이번 경기 국내 유일한 중계 ...
- 我是来围观的逊
- 반시간전
- Up
- Down
- Reply
- Favorite
-
"영비릉: 2024회계연도 영업수입 동기대비 8% 감소"영비릉은 2024회계연도 재무제보를 발표했다.2024 회계연도 매출은 149억5500만 유로로 전년 동기 대비 8% 감소했습니다.이익은 31억 500만 유로입니다.이익률은 ...
- 勇敢的树袋熊1
- 3 일전
- Up
- Down
- Reply
- Favorite
-
계면신문기자 장우발 4분기의 영업수입이 하락한후 텐센트음악은 다시 성장으로 돌아왔다. 11월 12일, 텐센트음악은 최신 재보를 발표했다.2024년 9월 30일까지 이 회사의 3분기 총수입은 70억 2천만 위안으로 전년 ...
- 勇敢的树袋熊1
- 그저께 15:27
- Up
- Down
- Reply
- Favorite
-
본사소식 (기자 원전새): 11월 14일, 다다그룹 (나스닥코드: DADA) 은 2024년 3분기 실적보고를 발표했다. 수치가 보여준데 따르면 고품질발전전략에 지속적으로 전념하고 사용자체험을 끊임없이 최적화하며 공급을 ...
- 家养宠物繁殖
- 어제 15:21
- Up
- Down
- Reply
- Favorite