첫 페이지 News 본문

On April 18th local time, the AI field received heavyweight news, and Meta officially released the long-awaited open source model Llama 3.
At the same time, Meta CEO Zuckerberg announced that based on the latest Llama 3 model, Meta's AI assistant has now covered a full range of applications such as Instagram, WhatsApp, Facebook, and has launched a separate website. Additionally, there is an image generator that can generate images based on natural language prompts.
Meta CEO Mark Zuckerberg stated in a video that the assistant can answer questions, create animations, and generate images.
Zuckerberg posted on Facebook: Big AI news today

Meta CEO Mark Zuckerberg said in the video: We believe that Meta AI is now the smartest artificial intelligence assistant you can freely use. Meta AI is built into the search boxes of WhatsApp, Instagram, Facebook, and Messenger applications, allowing users to easily ask questions that can be answered through new tools.
Zuckerberg stated that Meta's generative AI capabilities are the most powerful among free products.
On Facebook, Instagram, WhatsApp, and Messenger, users can now use Meta AI for searching without switching between applications:
When browsing the information flow, you can also directly ask Meta AI for more information from the post:
The gameplay brought by the image generator is more interesting, and the Image function brings real-time creation of images from text. The beta version of this feature will be launched on WhatsApp and Meta AI network experiences in the United States starting today.
When you start typing, you will see an image appearing, which changes with each additional letter input:
Meta states that Llama 3 outperforms industry-leading models in multiple key benchmark tests, achieving comprehensive leadership in code generation and other tasks. It is capable of complex reasoning, follows instructions more closely, visualizes ideas, and solves many subtle problems.
The main highlights of Llama 3 include:
Based on training with over 15T tokens, it is equivalent to more than 7 times the Llama 2 dataset;
Supporting 8K long text, the improved tokenizer has a vocabulary of 128K tokens, which can achieve better performance;
It has the most advanced performance among a large number of important benchmarks;
New ability categories, including enhanced reasoning and coding abilities;
The training efficiency is three times higher than Llama 2;
A new version of trust and security tools with Llama Guard 2, Code Shield, and CyberSec Eval 2.
The recently released 8B and 70B versions of the Llama 3 model have been used in Meta AI assistants and are also open source for developers, including pre trained and fine-tuning versions.
The latest release of Llama 3 models with 8B and 70B parameters can be said to be a significant leap in Llama 2. Due to improvements in pre training and post training, the pre training and instruction fine-tuning models released this time are the best models in the current size of 8B and 70B parameters. At the same time, the improvement of the post training process greatly reduces the model error rate, further improves consistency, and increases the diversity of model responses.
Llama 3 takes data and scale to new heights. Meta states that Llama 3 was trained on two customized 24K GPU clusters based on data exceeding 15T tokens - more than 7 times the size of the Llama 2 dataset, and 4 times the size of the code data. This has resulted in the most powerful Llama model to date, with Llama 3 supporting 8K context length, which is twice the capacity of Llama 2.
In addition, Meta has developed a new set of high-quality human assessment datasets. This evaluation set contains 1800 prompts, covering 12 key use cases: seeking advice, brainstorming, categorization, closed ended Q&A, coding, creative writing, extraction, character shaping, open ended Q&A, reasoning, rewriting, and summarizing. In order to prevent overfitting of Llama 3 on this evaluation set, Meta stated that their own team is also unable to access the dataset. The following figure shows the summary results of manual evaluation of these categories and prompts for Claude Sonnet, Mistral Medium, and GPT-3.5.
The following figure shows a comparison between the Llama 3 pre trained model and other models of the same scale, with the former exhibiting a SOTA level.
Managing large, high-quality training datasets is crucial for training the best language models. Meta has invested a significant amount of cost in pre training data. Llama 3 was pre trained using tokens exceeding 15T, all of which were collected from public sources. Overall, the training dataset of Llama 3 is more than seven times larger than the dataset used by Llama 2, and it contains more than four times more code. In preparation for the upcoming multilingual use cases, over 5% of the Llama 3 pre training dataset consists of high-quality non English data covering over 30 languages. However, the performance level of Llama 3 on these languages is not expected to be the same as English.
To ensure that Llama 3 receives training on the highest quality data, the research team developed a series of data filtering pipelines, including the use of heuristic filters (filters), NSFW filters, semantic duplicate data removal methods, and text classifiers to predict data quality.
The research team found that previous generations of Llama were very good at recognizing high-quality data, so Meta used Llama 2 to generate training data for the text quality classifier that supports Llama 3.
The research team also conducted extensive experiments to evaluate the optimal mixing method of data from different sources in the final pre training dataset, ultimately ensuring that Llama 3 performs well in various use cases (including daily problems, STEM, coding, historical knowledge, etc.).
Meta stated that the maximum Llama 3 parameters exceed 400B, and although these models are still in training, they will also be released in the coming months. New features include multimodal, multilingual dialogue ability, longer contextual windows, and stronger overall capabilities.
Meta hopes that Llama 3 can catch up with OpenAI's GPT-4. However, informed sources have revealed that as researchers have not yet started fine-tuning Llama 3, it has not been decided whether it will be a multimodal model. Fine tuning is the process by which developers provide additional data to existing models so that they can learn new information or tasks. Larger models typically provide higher quality responses, while smaller models tend to provide responses faster. It is reported that the official version of Llama 3 will be officially launched in July this year.
Meta also announced a new partnership with Alphabet's Google, which includes real-time search results in the assistant's response as a supplement to its existing partnership with Microsoft Bing. With this update, Meta AI Assistant is expanding to more than ten markets outside of the United States, including Australia, Canada, Singapore, Nigeria, and Pakistan. Cox said that Meta is "still working hard to do this in the right way in Europe.". Europe's privacy regulations are becoming stricter, and the upcoming artificial intelligence bill is also preparing to propose requirements such as disclosing model training data.
您需要登录后才可以回帖 登录 | Sign Up

本版积分规则

  • 11월 14일, 세계예선 아시아지역 제3단계 C조 제5라운드, 중국남자축구는 바레인남자축구와 원정경기를 가졌다.축구 국가대표팀은 바레인을 1-0으로 꺾고 예선 2연승을 거두었다. 특히 이번 경기 국내 유일한 중계 ...
    我是来围观的逊
    반시간전
    Up
    Down
    Reply
    Favorite
  • "영비릉: 2024회계연도 영업수입 동기대비 8% 감소"영비릉은 2024회계연도 재무제보를 발표했다.2024 회계연도 매출은 149억5500만 유로로 전년 동기 대비 8% 감소했습니다.이익은 31억 500만 유로입니다.이익률은 ...
    勇敢的树袋熊1
    3 일전
    Up
    Down
    Reply
    Favorite
  • 계면신문기자 장우발 4분기의 영업수입이 하락한후 텐센트음악은 다시 성장으로 돌아왔다. 11월 12일, 텐센트음악은 최신 재보를 발표했다.2024년 9월 30일까지 이 회사의 3분기 총수입은 70억 2천만 위안으로 전년 ...
    勇敢的树袋熊1
    그저께 15:27
    Up
    Down
    Reply
    Favorite
  • 본사소식 (기자 원전새): 11월 14일, 다다그룹 (나스닥코드: DADA) 은 2024년 3분기 실적보고를 발표했다. 수치가 보여준데 따르면 고품질발전전략에 지속적으로 전념하고 사용자체험을 끊임없이 최적화하며 공급을 ...
    家养宠物繁殖
    어제 15:21
    Up
    Down
    Reply
    Favorite
万小盈 新手上路
  • Follow

    0

  • Following

    0

  • Articles

    3