Takeaways
Today, we released our new Meta AI, one of the world’s leading free AI assistants built with Meta Llama 3, the next generation of our publicly available, state-of-the-art large language models. Thanks to our latest advances with Llama 3, Meta AI is smarter, faster, and more fun than ever before.
We are committed to developing AI responsibly and helping others do the same. That’s why we’re taking a series of steps so people can have enjoyable experiences when using these features and models, and sharing resources and tools to support developers and the open community.
Responsibility at different layers of the development process
We’re excited about the potential that generative AI technology can have for people who use Meta products, and for the broader ecosystem. We also want to make sure we’re developing and releasing this technology in a way that anticipates and works to reduce risk. To do this, we take steps to evaluate and address risks at each level of the AI development and deployment process. This includes incorporating protections in the process we use to design and release the Llama base model, supporting the developer ecosystem so they can build responsibly, and adopting the same best practices we expect of other developers when we develop and release our own generative AI features on Facebook, Instagram, WhatsApp, and Messenger.
As we explained when we released Llama 2, it’s important to be intentional in designing these mitigations because there are some measures that can only be effectively implemented by the model provider, and others that only work effectively when implemented by the developer as part of their specific application.
For these reasons, with Llama we take a system-centric approach that applies protections at every layer of the development stack. This includes taking a thoughtful approach to our training and tuning efforts and providing tools that make it easy for developers to implement models responsibly. In addition to maximizing the effectiveness of our responsible AI efforts, this approach aligns with our open innovation approach by giving developers more power to customize their products so they’re safer and benefit their users. The Responsible Use Guide is an important resource for developers that outlines considerations they should take to build their own products, which is why we followed its main steps when building Meta AI.
Responsibly building Llama 3 as a foundation model
We took several steps at the model level to develop a highly-capable and safe foundation model in Llama 3, including:
1. Addressing risks in training
The foundation of any model is the training process, through which the model learns both the language and information that it needs to operate. As a result, our approach started with a series of responsible AI mitigations in our training process. For example:
2. Safety evaluations and tuning
We adapted the pretrained model through a process called fine-tuning where we take additional steps to improve its performance in understanding and generating text conversations so it can be used for assistant-like chat applications.
During and after training, we conducted both automated and manual evaluations to understand our models’ performance in a series of risk areas like weapons, cyber attacks, and child exploitation. In each area, we performed additional work to limit the chance the model provides unwanted responses in these areas.
3. Lowering benign refusals
We’ve heard feedback from developers that Llama 2 would sometimes inadvertently refuse to answer innocuous prompts. Large language models tend to over-generalize, and we don’t intend for it to refuse to answer prompts like “How do I kill a computer program?” even though we don’t want it to respond to prompts like “How do I kill my neighbor?”
4. Model transparency
As with Llama 2, we’re publishing a model card that includes detailed information on Llama 3’s model architecture, parameters, and pretrained evaluations. The model card also provides information about the capabilities and limitations of the models.
Over the coming months, we’ll release additional Llama 3 models with new capabilities including multimodality, the ability to converse in multiple languages, and stronger overall capabilities. Our general approach of open sourcing our Llama 3 models is something we remain committed to. We’re currently training a 400B parameter model—and any final decision on when, whether, and how to open source will be taken following safety evaluations we will be running in the coming months.
How we built Meta AI as a responsible developer
We built the new Meta AI on top of Llama 3, just as we envision that Llama 3 will empower developers to expand the existing ecosystem of Llama-based products and services. As we describe in our Responsible Use Guide, we took additional steps at the different stages of product development and deployment to build Meta AI on top of the foundation model, just as any developer would use Llama 3 to build their own product.
In addition to the mitigations that we adopted within Llama 3, a developer needs to adopt additional mitigations to ensure the model can operate properly in the context of their specific AI system and use case. For Meta AI, the use case is a safe, helpful assistant available to people for free directly in our apps. We designed it to help people get things done like brainstorming and overcoming writer’s block, or connecting with friends to discover new places and adventures.
Since the launch of Meta AI last year, we’ve consistently updated and improved the experience and we’re continuing to make it even better. For example:
1. We improved Meta AI’s responses to peoples’ prompts and questions.
2. We taught the Meta AI model specific instructions and responses to make it a more helpful AI assistant.
3. We evaluated Meta AI’s performance against benchmarks and using human experts.
4. We applied safeguards at the prompt and response level.
5. We’ve built feedback tools within Meta AI.
Transparency is critical to help people understand this new technology and become comfortable with it. When someone interacts with Meta AI, we tell them it’s AI technology so they can choose whether they want to continue using it. We share information within the features themselves to help people understand that AI might return inaccurate or inappropriate outputs, which is the same for all generative AI systems. In chats with Meta AI, people can access additional information about how it generates content, the limitations of AI, and how the data they have shared with Meta AI is used.
We also include visible markers on photorealistic images generated by Meta AI so people know the content was created with AI. In May, we will begin labeling video, audio, and image content that people post on our apps as “Made with AI” when we detect industry standard AI image indicators or when people disclose that they’re uploading AI-generated content.
How developers can build responsibly with Llama 3
Meta AI is just one of many features and products that will be built with Llama 3, and we’re releasing different models in 8B and 70B sizes so developers can use the best version for them. We’re providing an instruction-tuned model that is specialized for chatbot applications and a pretrained model for developers with specific use cases that would benefit from custom policies.
In addition to the Responsible Use Guide, we’re providing open source tools that make it even easier for developers to customize Llama 3 and deploy generative AI-powered experiences.
Meta’s open approach to supporting the ecosystem
For more than a decade, Meta has been at the forefront of responsible open source in AI, and we believe that an open approach to AI leads to better, safer products, faster innovation, and a larger market. We’ve seen people using Llama 2 in new and innovative ways since it was released in July 2023—like Yale's Meditron LLM that’s helping medical professionals with decision-making and the Mayo Clinic’s tool that helps radiologists create clinically accurate summaries of their patients’ scans. Llama 3 has the potential to make these tools and experiences even better.
“The upcoming improvements in the reasoning capabilities of Llama 3 are important to any application, but especially in the medical domain, where trust depends quite a lot on the transparency of the decision-making process. Breaking down a decision/prediction into a set of logical steps is often how humans explain their actions and this kind of interpretability is expected from clinical decision support tools. Llama 2 not only enabled us to make Meditron, it also set a precedent for the potential impact of open-source foundation models in general. We are excited about Llama 3 for the example it sets in industry on the social value of open models.” —Prof Mary-Anne Hartley (Ph.D. MD, MPH), Director of the Laboratory for Intelligent Global Health and Humanitarian Response Technologies based jointly at Yale School of Medicine and EPFL School of Computer Science
Open source software is typically safer and more secure due to ongoing feedback, scrutiny, development, and mitigations from the community. Deploying AI safely is a shared responsibility of everyone in the ecosystem, which is why we’ve collaborated for many years with organizations that are working to build safe and trustworthy AI. For example, we’ve been working with MLCommons and a global set of partners to create responsibility benchmarks in ways that benefit the entire open source community. We co-founded the AI Alliance, a coalition of companies, academics, advocates, and governments working to develop tools that enable an open and safe AI ecosystem. We also recently released the findings from a Community Forum in partnership with Stanford and the Behavioral Insights Team so companies, researchers, and governments can make decisions based on input from people around the world about what’s important to them when it comes to generative AI chatbots.
We are collaborating with governments around the world to create a solid foundation for AI advancements to be secure, fair, and reliable. We eagerly await the progress on safety evaluation and research from national safety institutes including those in the United States and United Kingdom, particularly as they focus on establishing standardized threat models and evaluations throughout the AI development process. This will help measure risks quantitatively and consistently so risk thresholds can be set. The results of these efforts will guide companies like Meta in measuring and addressing risks, and deciding how and whether to release models.
As technologies continue to evolve, we look forward to improving these features and models in the months and years to come. And we look forward to helping people build, create, and connect in new and exciting ways.
Our latest updates delivered to your inbox
Subscribe to our newsletter to keep up with Meta AI news, events, research breakthroughs, and more.
Join us in the pursuit of what’s possible with AI.
Foundational models
Latest news
Foundational models