Common Objects in 3D (CO3D) is a dataset designed for learning category-specific 3D reconstruction and new-view synthesis using multi-view images of common object categories. The dataset has been introduced in our ICCV 2021 Paper
Learning to reconstruct the 3D structure of object categories has mainly been explored using only synthetic datasets due to the unavailability of real data. CO3D facilitates advances in this field by providing a large-scale dataset composed of real multi-view images of object categories annotated with camera poses and ground-truth 3D point clouds.
The CO3D dataset contains a total of 1.5 million frames from nearly 19,000 videos capturing objects from 50 MS-COCO categories. As such, it surpasses alternatives in terms of both the number of categories and objects. The dataset is suitable for learning category-specific 3D reconstruction and new-view synthesis methods, such as the seminal NeRF.
Videos: 18,619
Categories: 50 MS-COCO
Camera-annotated frames: 1.5 million
Point-cloud-annotated videos: 5,625
Download the dataset here.
Clone the github repository.
Read the README.md that describes how to visualize and evaluate on the dataset.
Foundational models
Latest news
Foundational models