SceneNet: Understanding Real World Indoor Scenes With Synthetic Data
MetadataShow full item record
Handa, A., Patraucean, V., Badrinarayanan, V., Stent, S., & Cipolla, R. SceneNet: Understanding Real World Indoor Scenes With Synthetic Data. https://doi.org/10.17863/CAM.26486
Scene understanding is a prerequisite to many high level tasks for any automated intelligent machine operating in real world environments. Recent attempts with supervised learning have shown promise in this direction but also high- lighted the need for enormous quantity of supervised data — performance increases in proportion to the amount of data used. However, this quickly becomes prohibitive when con- sidering the manual labour needed to collect such data. In this work, we focus our attention on depth based semantic per-pixel labelling as a scene understanding problem and show the potential of computer graphics to generate virtu- ally unlimited labelled data from synthetic 3D scenes. By carefully synthesizing training data with appropriate noise models we show comparable performance to state-of-the- art RGBD systems on NYUv2 dataset despite using only depth data as input and set a benchmark on depth-based segmentation on SUN RGB-D dataset.
This record's DOI: https://doi.org/10.17863/CAM.26486
This record's URL: https://www.repository.cam.ac.uk/handle/1810/279106