Abstract
This paper proposes a new framework for estimating the Manhattan Frame (MF) of an indoor scene from a single RGB-D image. Our technique formulates this problem as the estimation of a rotation matrix that best aligns the normals of the captured scene to a canonical world axes. By introducing sparsity constraints, our method can simultaneously estimate the scene MF, the surfaces in the scene that are best aligned to one of three coordinate axes, and the outlier surfaces that do not align with any of the axes. To test our approach, we contribute a new set of annotations to determine ground truth MFs in each image of the popular NYUv2 dataset. We use this new benchmark to experimentally demonstrate that our method is more accurate, faster, more reliable and more robust than the methods used in the literature. We further motivate our technique by showing how it can be used to address the RGB-D SLAM problem in indoor scenes by incorporating it into and improving the performance of a popular RGB-D SLAM method.
Original language | English (US) |
---|---|
Title of host publication | IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015 |
Publisher | IEEE Computer Society |
Pages | 3772-3780 |
Number of pages | 9 |
ISBN (Electronic) | 9781467369640 |
DOIs | |
State | Published - Oct 14 2015 |
Event | IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015 - Boston, United States Duration: Jun 7 2015 → Jun 12 2015 |
Publication series
Name | Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition |
---|---|
Volume | 07-12-June-2015 |
ISSN (Print) | 1063-6919 |
Conference
Conference | IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2015 |
---|---|
Country/Territory | United States |
City | Boston |
Period | 06/7/15 → 06/12/15 |
Bibliographical note
Publisher Copyright:© 2015 IEEE.
ASJC Scopus subject areas
- Software
- Computer Vision and Pattern Recognition