RESEARCH

COMPUTER VISION

SAM 3: Segment Anything with Concepts

November 19, 2025

Abstract

We present Segment Anything Model (SAM) 3, a unified model that detects, segments, and tracks objects in images and videos based on concept prompts, which we define as either short noun phrases (e.g., “yellow school bus”), image exemplars, or a combination of both. Promptable Concept Segmentation (PCS) takes such prompts and returns segmentation masks and unique identities for all matching object instances. To advance PCS, we build a scalable data engine that produces a high-quality dataset with 4M unique concept labels, including hard negatives, across images and videos. Our model consists of an image-level detector and a memory-based video tracker that share a single backbone. Recognition and localization are decoupled with a presence head, which boosts detection accuracy. SAM 3 delivers a 2× gain over existing systems in both image and video PCS, and improves previous SAM capabilities on visual segmentation tasks. We open source SAM 3 along with our new Segment Anything with Concepts (SA-Co) benchmark for promptable concept segmentation.

Download the Paper

AUTHORS

Written by

Nicolas Carion

Laura Gustafson

Yuan-Ting Hu

Shoubhik Debnath

Ronghang Hu

Didac Suris Coll-Vinent

Chaitanya Ryali

Kalyan Vasudev Alwala

Haitham Khedr

Andrew Huang

Jie Lei

Tengyu Ma

Baishan Guo

Arpit Kalla

Markus Marks

Joseph Greer

Meng Wang

Peize Sun

Roman Rädle

Triantafyllos Afouras

Effrosyni Mavroudi

Katherine Xu

Tsung-Han Wu

Yu Zhou

Liliane Momeni

Rishi Hazra

Shuangrui Ding

Sagar Vaze

Francois Porcher

Feng Li

Siyuan Li

Aishwarya Kamath

Ho Kei Cheng

Piotr Dollar

Nikhila Ravi

Kate Saenko

Pengchuan Zhang

Christoph Feichtenhofer

Publisher

arxiv

Research Topics

Computer Vision

Related Publications

November 19, 2025

COMPUTER VISION

SAM 3D: 3Dfy Anything in Images

SAM 3D Team, Xingyu Chen, Fu-Jen Chu, Pierre Gleize, Kevin J Liang, Alexander Sax, Hao Tang, Weiyao Wang, Michelle Guo, Thibaut Hardin, Xiang Li, Aohan Lin, Jiawei Liu, Ziqi Ma, Anushka Sagar, Bowen Song, Xiaodong Wang, Jianing Yang, Bowen Zhang, Piotr Dollar, Georgia Gkioxari, Matt Feiszli, Jitendra Malik

November 19, 2025

November 19, 2025

COMPUTER VISION

SAM 3D Body: Robust Full-Body Human Mesh Recovery

Xitong Yang, Devansh Kukreja, Don Pinkus, Anushka Sagar, Taosha Fan, Jinhyung Park, Soyong Shin, Jinkun Cao, Jiawei Liu, Nicolas Ugrinovic, Matt Feiszli, Jitendra Malik, Piotr Dollar, Kris Kitani

November 19, 2025

November 18, 2025

RESEARCH

CORE MACHINE LEARNING

Souper-Model: How Simple Arithmetic Unlocks State-of-the-Art LLM Performance

Shalini Maiti *, Amar Budhiraja *, Bhavul Gauri, Gaurav Chaurasia, Anton Protopopov, Alexis Audran-Reiss, Michael Slater, Despoina Magka, Tatiana Shavrina, Roberta Raileanu, Yoram Bachrach, * Equal authorship

November 18, 2025

November 11, 2025

COMPUTER VISION

SYSTEMS RESEARCH

CATransformers: Carbon Aware Transformers Through Joint Model-Hardware Optimization

Irene Wang, Mostafa Elhouishi, Ekin Sumbul, Samuel Hsia, Daniel Jiang, Newsha Ardalani, Divya Mahajan, Carole-Jean Wu, Bilge Acun

November 11, 2025

Help Us Pioneer The Future of AI

We share our open source frameworks, tools, libraries, and models for everything from research exploration to large-scale production deployment.