Responsible AI
Connect 2024: The responsible approach we’re taking to generative AI
September 25, 2024
5 minute read

Today at Connect 2024, we shared updates for Meta AI features and released Llama 3.2, a collection of models that includes new vision capabilities as well as lightweight models that can fit on mobile devices. With the rapidly evolving AI landscape, we recognize the importance of sharing our responsibility and safety approach with everyone—whether you’re a developer building with Llama or you’re using Meta AI experiences to help learn, create, and connect with the things and people that matter to you.

Our approach to safety for Llama 3.2

Llama 3.2 is built on the same foundation as our recent 3.1 release, which includes a multilayered safety approach, tools for developers, and extensive testing and evaluations work. Across all of our 3.2 models, we applied pre-training data mitigations to ensure a base level of safety. We also did thorough risk assessments of our fine-tuned Llame 3.2 models, tested the models’ performance, and fine-tuned features to help ensure the models are safe and reliable. This work includes conducting red-teaming exercises, such as cybersecurity and adversarial machine learning, and safety evaluations for fine-tuned models. Because Llama 3.2 models now include vision capabilities, we added additional measures to our safety program:

  • Safeguards for new vision capabilities:
    In Llama 3.1, we evaluated each fine-tuned model against a range of risks, including in areas like violent crime, child sexual exploitation, CBRNE (chemical, biological, radiological, nuclear, and explosives), and privacy. For Llama 3.2, we extended this work to evaluate possible risks associated with vision capabilities. We put each fine-tuned model through a variety of image reasoning tests including scaled evaluations and extensive red-teaming for possible image-input scenarios. These measures help to protect against potential bad actors.
  • New tools for developers—Llama Guard Vision and Llama Guard Update:
    Along with today’s Llama 3.2 release, we’re introducing Llama Guard Vision, which builds on the safety tools we previously released with Llama 3.1. Similar to the way Llama Guard provides moderation of text inputs and outputs, Llama Guard Vision is designed to support Llama 3.2’s new image understanding capability by detecting potentially problematic text and image input prompts, as well as text output responses. We’re also releasing a new, lightweight 1B version of Llama Guard that can be optimized to fit on devices while still achieving high performance. As with past releases, we’ve updated our Responsible Use Guide and released released model cardsto help developers learn more about using Llama Guard Vision and our smaller 1B Llama Guard.
  • Building industry standards:
    Both our existing Llama Guard and newly released Llama Guard Vision incorporate the hazard taxonomy developed by MLCommons. Through our ongoing collaboration with MLCommons, we’re working alongside researchers, security experts, and industry peers to create a set of third-party tools for evaluating and mitigating a wide-range of possible risks. This maximizes the power of the AI community to develop the safest and most useful models.

Making AI openly available will meaningfully improve the lives of people around the world. We’re committed to not only deploying our models openly, but also collaborating on the responsible development and use of AI. A collective, global effort is essential to safe AI innovation. By making our models open source, we empower the community to continually review, help issue-spot, and improve Llama models, making them more secure. We seek expertise from academics and researchers who use our models and civil society groups to make sure our approach is inclusive. We’ll continue to have these important conversations and take a collaborative approach to help ensure that we’re innovating responsibly and delivering AI experiences that work well for everyone.


Safeguards and resources for Meta AI features


We’ve developed our new image and voice features with safety and privacy in mind. In regions where people are able to upload images to Meta AI, we’ve taken steps to prevent Meta AI from being used to identify people in those images, such as safety-tuning to help detect prompts that ask Meta AI to identify who is in an image and output filtering to help prevent responses. We built safeguards to help protect against image edits resulting in harmful or inappropriate content. Because Meta AI now supports voice, we expanded our deletion controls so that voice transcriptions from Meta AI chat history can be deleted at any time.

People should know when they’re seeing and interacting with AI-related content. When people first begin to use our generative AI features, we have introductory, in-product experiences to help explain how to best use them. Images generated or edited by Meta AI’s Imagine feature also include visual watermarks to make it clear that these images have been generated with AI. Invisible watermarks and metadata are embedded within the image files as additional layers of transparency. And, we recently joined the C2PA steering committee to continue this important work.

We know people may have questions about the information generative AI features are trained on and how that information is used. We use things like public posts and commentsfrom Instagram and Facebook to develop and improve our AI products and tools, in addition to other kinds of data, like publicly available and licensed information from across the internet. We use the information that people share when interacting with our generative AI features, like Meta AI. We’re also clear about what we don’t use—for example, we did not train our Llama 3.2 models using posts or comments with an audience other than public. You can learn more about the data we collect and how we use your information by visiting our guide about AI at Meta in our Privacy Center as well as the Meta AI Terms of Service and our Privacy Policy.


Share:

Our latest updates delivered to your inbox

Subscribe to our newsletter to keep up with Meta AI news, events, research breakthroughs, and more.

Join us in the pursuit of what’s possible with AI.

Related Posts
FEATURED
Open Source
Llama 3.2: Revolutionizing edge AI and vision with open, customizable models
September 25, 2024
With 10x growth since 2023, Llama is the leading engine of AI innovation
August 29, 2024
Open Source
Generate an entire app from a prompt using Together AI’s LlamaCoder
September 18, 2024