查看原文
其他

Meta Llama 3真的来了!原文中英对照详解!

Amandahaha 慢达快语
2024-08-22

Hello!我是你的老朋友阿曼达。好久不见!

千呼万唤,Llama 3终于来了!

原文链接🔗:https://ai.meta.com/blog/meta-llama-3/

翻译:AI

Takeaways:

  • Today, we’re introducing Meta Llama 3, the next generation of our state-of-the-art open source large language model.

今天,我们将介绍Meta-Lama 3,这是我们最先进的开源大型语言模型的下一代。

  • Llama 3 models will soon be available on AWS, Databricks, Google Cloud, Hugging Face, Kaggle, IBM WatsonX, Microsoft Azure, NVIDIA NIM, and Snowflake, and with support from hardware platforms offered by AMD, AWS, Dell, Intel, NVIDIA, and Qualcomm.

Llama 3机型将很快在AWS、Databricks、Google Cloud、Hugging Face、Kaggle、IBM WatsonX、Microsoft Azure、NVIDIA NIM和Snowflake上推出,并得到AMD、AWS、Dell、Intel、NVIDIA和高通提供的硬件平台的支持。

  • We’re dedicated to developing Llama 3 in a responsible way, and we’re offering various resources to help others use it responsibly as well. This includes introducing new trust and safety tools with Llama Guard 2, Code Shield, and CyberSec Eval 2.

我们致力于以负责任的方式开发Llama 3,并提供各种资源帮助其他人负责任地使用它。这包括通过Llama Guard 2、Code Shield和CyberSec Eval 2引入新的信任和安全工具。

  • In the coming months, we expect to introduce new capabilities, longer context windows, additional model sizes, and enhanced performance, and we’ll share the Llama 3 research paper.

在接下来的几个月里,我们预计将推出新功能、更长的上下文窗口、额外的模型大小和增强的性能,我们将分享Llama 3的研究论文。

  • Meta AI, built with Llama 3 technology, is now one of the world’s leading AI assistants that can boost your intelligence and lighten your load—helping you learn, get things done, create content, and connect to make the most out of every moment. You can try Meta AI here.

Meta AI采用Llama 3技术构建,现在是世界领先的人工智能助手之一,可以提高你的智力,减轻你的负担——帮助你学习、完成任务、创建内容,并建立联系,充分利用每一刻。你可以在这里尝试Meta AI。


原文:

Today, we’re excited to share the first two models of the next generation of Llama, Meta Llama 3, available for broad use. This release features pretrained and instruction-fine-tuned language models with 8B and 70B parameters that can support a broad range of use cases. This next generation of Llama demonstrates state-of-the-art performance on a wide range of industry benchmarks and offers new capabilities, including improved reasoning. We believe these are the best open source models of their class, period. In support of our longstanding open approach, we’re putting Llama 3 in the hands of the community. We want to kickstart the next wave of innovation in AI across the stack—from applications to developer tools to evals to inference optimizations and more. We can’t wait to see what you build and look forward to your feedback.

今天,我们很高兴与大家分享下一代Llama的前两款车型Meta Llama 3,可供广泛使用。此版本的特点是预训练和指令微调的语言模型,具有8B和70B参数,可以支持广泛的用例。下一代Llama在广泛的行业基准测试中展示了最先进的性能,并提供了新的功能,包括改进的推理。我们相信这些是同类中最好的开源模型。为了支持我们长期以来的开放方法,我们将Llama 3交给社区。我们希望在整个堆栈中启动人工智能的下一波创新——从应用程序到开发工具,从评估到推理优化等等。我们迫不及待地想看看您构建了什么,并期待您的反馈。

Our goals for Llama 3

我们对Llama 3的目标

With Llama 3, we set out to build the best open models that are on par with the best proprietary models available today. We wanted to address developer feedback to increase the overall helpfulness of Llama 3 and are doing so while continuing to play a leading role on responsible use and deployment of LLMs. We are embracing the open source ethos of releasing early and often to enable the community to get access to these models while they are still in development. The text-based models we are releasing today are the first in the Llama 3 collection of models. Our goal in the near future is to make Llama 3 multilingual and multimodal, have longer context, and continue to improve overall performance across core LLM capabilities such as reasoning and coding.

有了Llama 3,我们开始构建与当今最佳专有模型不相上下的最佳开放模型。我们希望解决开发人员的反馈问题,以提高Llama 3的整体实用性,同时继续在LLM的负责任使用和部署方面发挥主导作用。我们正在接受早期和经常发布的开源精神,以使社区能够在这些模型仍在开发中时访问它们。我们今天发布的基于文本的模型是Llama 3模型系列中的第一个。在不久的将来,我们的目标是使Llama 3成为多语言和多模式的,具有更长的上下文,并继续提高推理和编码等核心LLM功能的整体性能。

State-of-the-art performance

最先进的表现

Our new 8B and 70B parameter Llama 3 models are a major leap over Llama 2 and establish a new state-of-the-art for LLM models at those scales. Thanks to improvements in pretraining and post-training, our pretrained and instruction-fine-tuned models are the best models existing today at the 8B and 70B parameter scale. Improvements in our post-training procedures substantially reduced false refusal rates, improved alignment, and increased diversity in model responses. We also saw greatly improved capabilities like reasoning, code generation, and instruction following making Llama 3 more steerable.

我们新的8B和70B参数Llama 3模型是对Llama 2的重大飞跃,并为这些规模的LLM模型建立了新的最先进的技术。由于在预训练和后训练方面的改进,我们的预训练和教学微调模型是目前8B和70B参数范围内最好的模型。我们训练后程序的改进大大降低了错误拒绝率,改善了一致性,并增加了模型响应的多样性。我们还看到,推理、代码生成和指令跟随等功能得到了极大的改进,使Llama 3更加可控。

*Please see evaluation details for setting and parameters with which these evaluations are calculated. *有关计算这些评估的设置和参数,请参阅评估详细信息。

In the development of Llama 3, we looked at model performance on standard benchmarks and also sought to optimize for performance for real-world scenarios. To this end, we developed a new high-quality human evaluation set. This evaluation set contains 1,800 prompts that cover 12 key use cases: asking for advice, brainstorming, classification, closed question answering, coding, creative writing, extraction, inhabiting a character/persona, open question answering, reasoning, rewriting, and summarization. To prevent accidental overfitting of our models on this evaluation set, even our own modeling teams do not have access to it. The chart below shows aggregated results of our human evaluations across of these categories and prompts against Claude Sonnet, Mistral Medium, and GPT-3.5.

在Llama 3的开发过程中,我们研究了标准基准上的模型性能,并试图优化真实场景的性能。为此,我们开发了一套新的高质量人类评估集。该评估集包含1800个提示,涵盖12个关键用例:征求建议、头脑风暴、分类、封闭式问题回答、编码、创造性写作、提取、角色/角色、开放式问题回答,推理、重写和总结。为了防止我们的模型在该评估集上意外过度拟合,即使是我们自己的建模团队也无法访问它。下表显示了我们对这些类别的人类评估的汇总结果,并针对Claude Sonnet、Mistral Medium和GPT-3.5进行提示。

Preference rankings by human annotators based on this evaluation set highlight the strong performance of our 70B instruction-following model compared to competing models of comparable size in real-world scenarios.

人类注释者基于该评估集进行的偏好排名突出了我们的70B指令遵循模型在现实世界场景中与同等规模的竞争模型相比的强大性能。

Our pretrained model also establishes a new state-of-the-art for LLM models at those scales.

我们的预训练模型也为LLM模型在这些尺度上建立了一个新的最先进的技术。

*Please see evaluation details for setting and parameters with which these evaluations are calculated. *有关计算这些评估的设置和参数,请参阅评估详细信息。

To develop a great language model, we believe it’s important to innovate, scale, and optimize for simplicity. We adopted this design philosophy throughout the Llama 3 project with a focus on four key ingredients: the model architecture, the pretraining data, scaling up pretraining, and instruction fine-tuning. 为了开发一个优秀的语言模型,我们认为创新、扩展和优化以实现简单性是很重要的。我们在整个Llama 3项目中采用了这一设计理念,重点关注四个关键要素:模型架构、预训练数据、放大预训练和指令微调。

Model architecture

模型架构

In line with our design philosophy, we opted for a relatively standard decoder-only transformer architecture in Llama 3. Compared to Llama 2, we made several key improvements. Llama 3 uses a tokenizer with a vocabulary of 128K tokens that encodes language much more efficiently, which leads to substantially improved model performance. To improve the inference efficiency of Llama 3 models, we’ve adopted grouped query attention (GQA) across both the 8B and 70B sizes. We trained the models on sequences of 8,192 tokens, using a mask to ensure self-attention does not cross document boundaries.

根据我们的设计理念,我们在Llama 3中选择了相对标准的仅限解码器的转换器架构。与Llama 2相比,我们做了几个关键改进。Llama 3使用了一个具有128K标记词汇表的标记器,该标记器对语言进行了更有效的编码,从而大大提高了模型性能。为了提高Llama 3模型的推理效率,我们在8B和70B大小上都采用了分组查询注意力(GQA)。我们在8192个标记的序列上训练模型,使用掩码来确保自我关注不会跨越文档边界。

Training data

训练数据

To train the best language model, the curation of a large, high-quality training dataset is paramount. In line with our design principles, we invested heavily in pretraining data. Llama 3 is pretrained on over 15T tokens that were all collected from publicly available sources. Our training dataset is seven times larger than that used for Llama 2, and it includes four times more code. To prepare for upcoming multilingual use cases, over 5% of the Llama 3 pretraining dataset consists of high-quality non-English data that covers over 30 languages. However, we do not expect the same level of performance in these languages as in English.

为了训练最好的语言模型,管理一个大的、高质量的训练数据集是至关重要的。根据我们的设计原则,我们在数据预训练方面投入了大量资金。Llama 3在超过15T的代币上进行了预训练,这些代币都是从公开来源收集的。我们的训练数据集比用于Llama 2的数据集大7倍,并且它包含的代码多4倍。为了准备即将到来的多语言用例,Llama 3预训练数据集的5%以上由覆盖30多种语言的高质量非英语数据组成。然而,我们并不期望在这些语言中有与英语相同的表现水平。

To ensure Llama 3 is trained on data of the highest quality, we developed a series of data-filtering pipelines. These pipelines include using heuristic filters, NSFW filters, semantic deduplication approaches, and text classifiers to predict data quality. We found that previous generations of Llama are surprisingly good at identifying high-quality data, hence we used Llama 2 to generate the training data for the text-quality classifiers that are powering Llama 3.

为了确保Llama 3在最高质量的数据上进行训练,我们开发了一系列数据过滤管道。这些管道包括使用启发式过滤器、NSFW过滤器、语义重复数据消除方法和文本分类器来预测数据质量。我们发现,前几代的Llama在识别高质量数据方面出奇地出色,因此我们使用Llama 2为支持Llama 3的文本质量分类器生成训练数据。

We also performed extensive experiments to evaluate the best ways of mixing data from different sources in our final pretraining dataset. These experiments enabled us to select a data mix that ensures that Llama 3 performs well across use cases including trivia questions, STEM, coding, historical knowledge, etc.

我们还进行了广泛的实验,以评估在我们的最终预训练数据集中混合来自不同来源的数据的最佳方式。这些实验使我们能够选择一种数据组合,确保Llama 3在用例中表现良好,包括琐事问题、STEM、编码、历史知识等。

Scaling up pretraining

扩大预训练

To effectively leverage our pretraining data in Llama 3 models, we put substantial effort into scaling up pretraining. Specifically, we have developed a series of detailed scaling laws for downstream benchmark evaluations. These scaling laws enable us to select an optimal data mix and to make informed decisions on how to best use our training compute. Importantly, scaling laws allow us to predict the performance of our largest models on key tasks (for example, code generation as evaluated on the HumanEval benchmark—see above) before we actually train the models. This helps us ensure strong performance of our final models across a variety of use cases and capabilities.

为了有效地利用我们在Llama 3模型中的预训练数据,我们投入了大量精力来扩大预训练。具体而言,我们为下游基准评估制定了一系列详细的比例定律。这些缩放定律使我们能够选择最佳的数据组合,并就如何最好地使用我们的训练计算做出明智的决定。重要的是,缩放定律使我们能够在实际训练模型之前预测最大模型在关键任务上的性能(例如,在HumanEval基准上评估的代码生成——见上文)。这有助于我们确保最终模型在各种用例和功能中具有强大的性能。

We made several new observations on scaling behavior during the development of Llama 3. For example, while the Chinchilla-optimal amount of training compute for an 8B parameter model corresponds to ~200B tokens, we found that model performance continues to improve even after the model is trained on two orders of magnitude more data. Both our 8B and 70B parameter models continued to improve log-linearly after we trained them on up to 15T tokens. Larger models can match the performance of these smaller models with less training compute, but smaller models are generally preferred because they are much more efficient during inference.

在Llama 3的开发过程中,我们对缩放行为进行了一些新的观察。例如,虽然8B参数模型的Chinchilla最优训练计算量对应于约200B个令牌,但我们发现,即使在多训练了两个数量级的数据后,模型性能仍在继续提高。在我们对8B和70B参数模型进行了多达15T的令牌训练后,它们都继续线性地提高对数。较大的模型可以用较少的训练计算来匹配这些较小模型的性能,但较小的模型通常是优选的,因为它们在推理过程中效率高得多。

To train our largest Llama 3 models, we combined three types of parallelization: data parallelization, model parallelization, and pipeline parallelization. Our most efficient implementation achieves a compute utilization of over 400 TFLOPS per GPU when trained on 16K GPUs simultaneously. We performed training runs on two custom-built 24K GPU clusters. To maximize GPU uptime, we developed an advanced new training stack that automates error detection, handling, and maintenance. We also greatly improved our hardware reliability and detection mechanisms for silent data corruption, and we developed new scalable storage systems that reduce overheads of checkpointing and rollback. Those improvements resulted in an overall effective training time of more than 95%. Combined, these improvements increased the efficiency of Llama 3 training by ~three times compared to Llama 2.

为了训练我们最大的Llama 3模型,我们结合了三种类型的并行化:数据并行化、模型并行化和流水线并行化。当同时在16K GPU上进行训练时,我们最高效的实现实现了每个GPU超过400 TFLOPS的计算利用率。我们在两个定制的24K GPU集群上进行了训练运行。为了最大限度地延长GPU的正常运行时间,我们开发了一个先进的新训练堆栈,可以自动检测、处理和维护错误。我们还大大提高了硬件可靠性和无声数据损坏的检测机制,并开发了新的可扩展存储系统,减少了检查点和回滚的开销。这些改进使总的有效训练时间超过95%。综合起来,这些改进使Llama 3的训练效率比Llama 2提高了约三倍。

Instruction fine-tuning

指令微调介绍

To fully unlock the potential of our pretrained models in chat use cases, we innovated on our approach to instruction-tuning as well. Our approach to post-training is a combination of supervised fine-tuning (SFT), rejection sampling, proximal policy optimization (PPO), and direct policy optimization (DPO). The quality of the prompts that are used in SFT and the preference rankings that are used in PPO and DPO has an outsized influence on the performance of aligned models. Some of our biggest improvements in model quality came from carefully curating this data and performing multiple rounds of quality assurance on annotations provided by human annotators.

为了在聊天用例中充分释放我们预训练模型的潜力,我们还创新了指令调整方法。我们的后训练方法是监督微调(SFT)、拒绝采样、近端策略优化(PPO)和直接策略优化(DPO)的组合。SFT中使用的提示的质量以及PPO和DPO中使用的偏好排名对对齐模型的性能有着巨大的影响。我们在模型质量方面的一些最大改进来自于仔细管理这些数据,并对人工注释器提供的注释执行多轮质量保证。

Learning from preference rankings via PPO and DPO also greatly improved the performance of Llama 3 on reasoning and coding tasks. We found that if you ask a model a reasoning question that it struggles to answer, the model will sometimes produce the right reasoning trace: The model knows how to produce the right answer, but it does not know how to select it. Training on preference rankings enables the model to learn how to select it.

通过PPO和DPO从偏好排名中学习也大大提高了Llama 3在推理和编码任务上的性能。我们发现,如果你问一个模型一个很难回答的推理问题,模型有时会产生正确的推理痕迹:模型知道如何产生正确的答案,但不知道如何选择。偏好排名的训练使模型能够学习如何选择。

Building with Llama 3

与Llama 3一起开发建设

Our vision is to enable developers to customize Llama 3 to support relevant use cases and to make it easier to adopt best practices and improve the open ecosystem. With this release, we’re providing new trust and safety tools including updated components with both Llama Guard 2 and Cybersec Eval 2, and the introduction of Code Shield—an inference time guardrail for filtering insecure code produced by LLMs. 我们的愿景是使开发人员能够定制Llama 3,以支持相关用例,并使其更容易采用最佳实践和改进开放生态系统。通过此次发布,我们将提供新的信任和安全工具,包括Llama Guard 2和Cybersec Eval 2的更新组件,以及引入Code Shield——一种用于过滤LLM产生的不安全代码的推理时间护栏。

We’ve also co-developed Llama 3 with torchtune, the new PyTorch-native library for easily authoring, fine-tuning, and experimenting with LLMs. torchtune provides memory efficient and hackable training recipes written entirely in PyTorch. The library is integrated with popular platforms such as Hugging Face, Weights & Biases, and EleutherAI and even supports Executorch for enabling efficient inference to be run on a wide variety of mobile and edge devices. For everything from prompt engineering to using Llama 3 with LangChain we have a comprehensive getting started guide and takes you from downloading Llama 3 all the way to deployment at scale within your generative AI application.

我们还与torchtune共同开发了Llama 3,这是一个新的PyTorch原生库,用于轻松创作、微调和试验LLM。torchtune提供了完全用PyTorch编写的高效记忆和可破解的训练食谱。该库与流行的平台集成,如Hugging Face、Weights&Biases和EleutherAI,甚至支持Executiorch,使高效推理能够在各种移动和边缘设备上运行。从即时工程到将Llama 3与LangChain一起使用,我们都有一份全面的入门指南,带您从下载Llama 2一直到在生成的人工智能应用程序中大规模部署。

A system-level approach to responsibility

一种系统级的责任方法

We have designed Llama 3 models to be maximally helpful while ensuring an industry leading approach to responsibly deploying them. To achieve this, we have adopted a new, system-level approach to the responsible development and deployment of Llama. We envision Llama models as part of a broader system that puts the developer in the driver’s seat. Llama models will serve as a foundational piece of a system that developers design with their unique end goals in mind.

我们设计了Llama 3型号,以最大限度地提供帮助,同时确保采用行业领先的方法负责任地部署它们。为了实现这一目标,我们采用了一种新的系统级方法来负责任地开发和部署Llama。我们设想Llama模型是一个更广泛的系统的一部分,让开发者坐在驾驶座上。Llama模型将成为开发人员在设计时考虑到其独特最终目标的系统的基础部分。

Instruction fine-tuning also plays a major role in ensuring the safety of our models. Our instruction-fine-tuned models have been red-teamed (tested) for safety through internal and external efforts. Our red teaming approach leverages human experts and automation methods to generate adversarial prompts that try to elicit problematic responses. For instance, we apply comprehensive testing to assess risks of misuse related to Chemical, Biological, Cyber Security, and other risk areas. All of these efforts are iterative and used to inform safety fine-tuning of the models being released. You can read more about our efforts in the model card.

指令微调也在确保我们模型的安全方面发挥着重要作用。我们的指导微调模型已经通过内部和外部努力进行了安全性测试。我们的红队方法利用人类专家和自动化方法来生成对抗性提示,试图引发有问题的反应。例如,我们应用综合测试来评估与化学、生物、网络安全和其他风险领域相关的滥用风险。所有这些努力都是迭代的,并用于通知正在发布的模型的安全微调。您可以在模型卡中阅读更多关于我们努力的信息。

Llama Guard models are meant to be a foundation for prompt and response safety and can easily be fine-tuned to create a new taxonomy depending on application needs. As a starting point, the new Llama Guard 2 uses the recently announced MLCommons taxonomy, in an effort to support the emergence of industry standards in this important area. Additionally, CyberSecEval 2 expands on its predecessor by adding measures of an LLM’s propensity to allow for abuse of its code interpreter, offensive cybersecurity capabilities, and susceptibility to prompt injection attacks (learn more in our technical paper). Finally, we’re introducing Code Shield which adds support for inference-time filtering of insecure code produced by LLMs. This offers mitigation of risks around insecure code suggestions, code interpreter abuse prevention, and secure command execution.

Llama-Guard模型旨在成为提示和响应安全的基础,并且可以根据应用程序的需要轻松地进行微调以创建新的分类法。作为一个起点,新的Llama Guard 2使用了最近宣布的MLCommons分类法,以支持在这一重要领域出现行业标准。此外,CyberSecEval 2对其前身进行了扩展,增加了LLM倾向于允许滥用其代码解释器、攻击性网络安全能力和易受即时注入攻击的指标(在我们的技术论文中了解更多信息)。最后,我们将介绍CodeShield,它增加了对LLM生成的不安全代码的推理时过滤的支持。这提供了对不安全代码建议、代码解释器滥用预防和安全命令执行的风险的缓解。

With the speed at which the generative AI space is moving, we believe an open approach is an important way to bring the ecosystem together and mitigate these potential harms. As part of that, we’re updating our Responsible Use Guide (RUG) that provides a comprehensive guide to responsible development with LLMs. As we outlined in the RUG, we recommend that all inputs and outputs be checked and filtered in accordance with content guidelines appropriate to the application. Additionally, many cloud service providers offer content moderation APIs and other tools for responsible deployment, and we encourage developers to also consider using these options.

随着生成人工智能空间的发展速度,我们相信开放的方法是将生态系统整合在一起并减轻这些潜在危害的重要途径。作为其中的一部分,我们正在更新我们的负责任使用指南(RUG),该指南为LLM的负责任开发提供了全面的指南。正如我们在RUG中概述的那样,我们建议根据适用于应用程序的内容指南检查和过滤所有输入和输出。此外,许多云服务提供商为负责任的部署提供内容审核API和其他工具,我们鼓励开发人员也考虑使用这些选项。

Deploying Llama 3 at scale

大规模部署Llama 3

Llama 3 will soon be available on all major platforms including cloud providers, model API providers, and much more. Llama 3 will be everywhere.

Llama 3将很快在所有主要平台上推出,包括云提供商、API模型提供商等。Llama 3将无处不在。

Our benchmarks show the tokenizer offers improved token efficiency, yielding up to 15% fewer tokens compared to Llama 2. Also, Group Query Attention (GQA) now has been added to Llama 3 8B as well. As a result, we observed that despite the model having 1B more parameters compared to Llama 2 7B, the improved tokenizer efficiency and GQA contribute to maintaining the inference efficiency on par with Llama 2 7B.

我们的基准测试显示,代币化器提高了代币效率,与Llama 2相比,代币产量减少了15%。此外,Group Query Attention(GQA)现在也已添加到Llama 3 8B中。因此,我们观察到,尽管与Llama 2 7B相比,该模型具有1B更多的参数,但改进的标记器效率和GQA有助于将推理效率保持在与Llama2 7B相当的水平。

For examples of how to leverage all of these capabilities, check out Llama Recipes which contains all of our open source code that can be leveraged for everything from fine-tuning to deployment to model evaluation.

有关如何利用所有这些功能的示例,请查看Llama Recipes,它包含了我们所有的开源代码,可以用于从微调到部署再到模型评估的所有方面。

What’s next for Llama 3?

Llama 3的下一步是什么?

The Llama 3 8B and 70B models mark the beginning of what we plan to release for Llama 3. And there’s a lot more to come.

Llama3 8B和70B型号标志着我们计划为Llama 3发布的产品的开始。还有很多事情要做。

Our largest models are over 400B parameters and, while these models are still training, our team is excited about how they’re trending. Over the coming months, we’ll release multiple models with new capabilities including multimodality, the ability to converse in multiple languages, a much longer context window, and stronger overall capabilities. We will also publish a detailed research paper once we are done training Llama 3.

我们最大的模型参数超过400B,当这些模型仍在训练时,我们的团队对它们的趋势感到兴奋。在接下来的几个月里,我们将发布多个具有新功能的模型,包括多模态、用多种语言进行对话的能力、更长的上下文窗口和更强的整体功能。我们还将在训练完Llama 3后发表一篇详细的研究论文。

To give you a sneak preview for where these models are today as they continue training, we thought we could share some snapshots of how our largest LLM model is trending. Please note that this data is based on an early checkpoint of Llama 3 that is still training and these capabilities are not supported as part of the models released today.

为了让您在这些模型继续训练时预览它们的现状,我们想分享一些我们最大的LLM模型的趋势快照。请注意,这些数据是基于Llama 3的早期检查点,该检查点仍在训练中,今天发布的模型不支持这些功能。

*Please see evaluation details for setting and parameters with which these evaluations are calculated. *有关计算这些评估的设置和参数,请参阅评估详细信息。

We’re committed to the continued growth and development of an open AI ecosystem for releasing our models responsibly. We have long believed that openness leads to better, safer products, faster innovation, and a healthier overall market. This is good for Meta, and it is good for society.We’re taking a community-first approach with Llama 3, and starting today, these models are available on the leading cloud, hosting, and hardware platforms with many more to come.

我们致力于开放人工智能生态系统的持续增长和发展,以负责任地发布我们的模型。我们长期以来一直相信,开放会带来更好、更安全的产品、更快的创新和更健康的整体市场。这对Meta有好处,对社会也有好处。我们对Llama 3采取了社区优先的方法,从今天开始,这些模型可以在领先的云、主机和硬件平台上使用,还有更多。

Try Meta Llama 3 today

立即尝试Meta Llama 3

We’ve integrated our latest models into Meta AI, which we believe is the world’s leading AI assistant. It’s now built with Llama 3 technology and it’s available in more countries across our apps.

我们已经将我们的最新模型集成到Meta AI中,我们相信Meta AI是世界领先的人工智能助手。它现在使用Llama 3技术构建,并可在我们的应用程序中在更多国家/地区使用。

You can use Meta AI on Facebook, Instagram, WhatsApp, Messenger, and the web to get things done, learn, create, and connect with the things that matter to you. You can read more about the Meta AI experience here.

你可以在Facebook、Instagram、WhatsApp、Messenger和网络上使用Meta AI来完成任务、学习、创建并与对你重要的事情建立联系。你可以在这里阅读更多关于Meta AI体验的信息。

Visit the Llama 3 website to download the models and reference the Getting Started Guide for the latest list of all available platforms.

访问Llama 3网站下载型号,并参考《入门指南》了解所有可用平台的最新列表。

You’ll also soon be able to test multimodal Meta AI on our Ray-Ban Meta smart glasses.

您还将很快能够在我们的雷朋Meta智能眼镜上测试多模式Meta AI。

As always, we look forward to seeing all the amazing products and experiences you will build with Meta Llama 3.

一如既往,我们期待着看到您将使用Meta Llama 3打造的所有令人惊叹的产品和体验。


*end.


欢迎点赞,点在看,关注我的公众号,一起成长!
记得把公号加星标,否则可能看不到推送:)‍


宇宙无垠,新一年,继续探索。与时间做朋友,与自己和解。‍‍‍‍

欢迎各位朋友和我分享互动。Have a nice day!


继续滑动看下一个
慢达快语
向上滑动看下一个

您可能也对以下帖子感兴趣

文章有问题?点此查看未经处理的缓存