ScanNet: Richly-annotated 3D Reconstructions of Indoor Scenes
Angela Dai Angel X. Chang Manolis Savva Maciej Halber
Thomas Funkhouser Matthias Nießner
- 2018-06-11 : ScanNet v2 release.
- 2018-06-11 : ScanNet Benchmark Challenge available.
- 2018-02-11 : Browsing interface available.
- 2018-02-04 : We are part of the Robust Vision Challenge 2018 Workshop at CVPR 2018 in Salt Lake City. Participate to test out your algorithms and win prizes!
- 2017-09-30 : Data server code released
- 2017-02-01 : ScanNet initial release
ScanNet is an RGB-D video dataset containing 2.5 million views in more than 1500 scans, annotated with 3D camera poses, surface reconstructions, and instance-level semantic segmentations. To collect this data, we designed an easy-to-use and scalable RGB-D capture system that includes automated surface reconstruction and crowdsourced semantic annotation. We show that using this data helps achieve state-of-the-art performance on several 3D scene understanding tasks, including 3D object classification, semantic voxel labeling, and CAD model retrieval. More information can be found in our paper.
If you use the ScanNet data or code please cite:
@inproceedings{dai2017scannet,
title={ScanNet: Richly-annotated 3D Reconstructions of Indoor Scenes},
author={Dai, Angela and Chang, Angel X. and Savva, Manolis and Halber, Maciej and Funkhouser, Thomas and Nie{\ss}ner, Matthias},
booktitle = {Proc. Computer Vision and Pattern Recognition (CVPR), IEEE},
year = {2017}
}
See detailed changelog here.
- 2018-11-06 : ScanNet v2 release.
The ScanNet data is released under the ScanNet Terms of Use, and the code is released under the MIT license.
The ScanNet data can be browsed online.
- Query scans
- Browse scans grouped by scene type
Please visit our main project repository for more information and access to code, data, and trained models: https://github.com/ScanNet/ScanNet
Semantic voxel labeling of 3D scans in ScanNet using our 3D CNN architecture. Voxel colors indicate predicted or ground truth category.The ScanNet project is funded by Google Tango, Intel, NSF (IIS-1251217 and VEC 1539014/1539099), and a Stanford Graduate fellowship. We also thank Occipital for donating structure sensors and Nvidia for hardware donations, as well as support by the Max-Planck Center for Visual Computing and the Stanford CURIS program.
Toan Vuong, Joseph Chang, and Helen Jiang helped develop the mobile scanning app, and Hope Casey-Allen and Duc Nugyen developed early prototypes of the annotation interfaces. We thank Alex Sabia for scanning and verifying annotations, and Halle Pollack, Julian Massarani and Michael Fang fo checking annotations. Last but not least, we would like to thank all the other volunteers who helped with scanning and getting us access to scanning spaces, as well as all the AMT workers who annotated and gave us feedback on the interface design.
--- Last updated: 2018-02-11