AI RESEARCH

We are innovating in the open,
for a smarter, more connected world

We are innovating in the open, for a smarter, more connected world

With a decade of experience, the Meta FAIR team aims to achieve advanced machine intelligence (AMI) to accelerate the potential of future Meta products for the benefit of people.
random 1
NEW

Segment Anything 3

With SAM 3, you can use text and visual prompts to precisely detect, segment and track any object in any image or video.
SAM 3D
SAM 3D enables precise reconstruction and analysis of 3D people and objects, providing new opportunities for spatial understanding and applications.
SAM Audio
SAM Audio can accurately separate target and residual sounds from any audio or audiovisual source.
WHAT'S NEW
CORE LEARNING & REASONING

DINOv3

DINOv3 scales self-supervised learning (SSL) for images to produce our strongest universal vision backbones, enabling breakthrough performance across diverse domains.

DINOv3 scales self-supervised learning to train powerful, versatile model

Learn more
CORE LEARNING & REASONING

V-JEPA 2

The first world model trained on video that achieves state-of-the-art visual understanding and prediction.

Video Joint Embedding Predictive Architecture 2 (V-JEPA 2) is a self-supervised foundation world model

Learn more
COMMUNICATION & LANGUAGE

Seamless Interaction

Advancing AI research modeling of face-to-face dynamics, including expressive gestures, active listening, turn-taking and visual synchrony.

Audiovisual motion models compatible with 2D and 3D renderings, trained on the Seamless Interaction Dataset

Learn more
PERCEPTION

Segment Anything 2

SAM 2 is a segmentation model that enables fast, precise selection of any object in any video or image.

Advanced capabilities in object detection, segmentation and tracking

Learn more

More from Meta's FAIR Team

Meta Motivo thumbnail

Meta Motivo

A first-of-its-kind behavioral foundation model for embodied humanoid virtual agents.
Meta Motivo
Video Seal thumbnail

Video Seal

A state-of-the-art, open-source model for video watermarking.
Video Seal
Movie Gen thumbnail

Movie Gen

The most advanced family of media foundation AI models empowering immersive storytelling.
Movie Gen
Audiobox thumbnail

Audiobox

A foundation research model for audio generation that can generate voices and sound effects.
Audiobox
Seamless Communication thumbnail

Seamless Communication

AI research models that enable more natural, authentic communication across languages.
Seamless Communication
AI Chemistry thumbnail

AI Chemistry

Building AI systems solving most important chemistry and material problems for Meta and the world.
AI Chemistry
Show More

Try experimental demos

How Meta is applying cutting-edge AI research to real-world interactions

A state-of-the-art, open-source model for video watermarking
Meta Video Seal

Create video cutouts and effects with a few clicks

SAM 2
OUR RESEARCH

For researchers and developers

Meta FAIR is advancing research and delivering breakthroughs in a variety of areas.
Resources & tools
RESEARCH AREAS

We advance AI capabilities in expressive communication, social interaction and use of language. Through foundational research in natural language processing and multimodal AI, we develop systems that enable more natural, meaningful interactions between humans and machines.

We advance the fundamental capabilities needed for AI to understand and act within the physical and digital world. Through our research, we hope to unlock a wide variety of future agents that help humans do more throughout all aspects of their lives. From robots that can move around, interact with objects, to help accomplish household tasks, to wearable glasses that understand the real and digital world and support people throughout their daily tasks.

Our research focuses on aligning models and decisions with human intent and societal interests through deeper fundamental understanding and enhanced steerability and efficiency of AI models. The pillar is at the forefront of research on AI for science and AI for society.

We conduct fundamental research in pre-training methods and new architectural paradigms that enable foundation models to learn and reason with agility and efficiency across novel downstream challenges. Our work expands the frontier of approaches such as world models, non-autoregressive architectures, and memory-augmented models to unlock new capabilities in adaptive intelligence.

We develop code world models as foundational models for code and agents, and advance methods to do reinforcement learning with execution feedback. We research how to do much more efficient architectures for code world models, latent space reasoning, and grounded reasoning and planning with world models. We develop various agents, e.g. AI research agents to help our own research, and upstream our agents’ needs to our foundational models.

The north star goal of our Perception research teams is to enable general AI systems to perceive the visual world to inform action, communication, and generation. To achieve this goal, we're developing next generation perception models capable of understanding images and videos not as pixels, but as a capture of visual entities like people, objects, activities, their spatial and temporal relationships.

Llama 4 our latest large language model

Our latest models bring mixture-of-experts architecture, native multimodal capabilities, near limitless context windows, and billion-scale performance.

Download the latest model
Working in the office

Join us in the pursuit of what’s possible with AI

Open positions