Translator Disclaimer
7 March 2014 Texture mapping 3D models of indoor environments with noisy camera poses
Author Affiliations +
Proceedings Volume 9020, Computational Imaging XII; 90200V (2014)
Event: IS&T/SPIE Electronic Imaging, 2014, San Francisco, California, United States
Automated 3D modeling of building interiors is used in applications such as virtual reality and environment mapping. Texturing these models allows for photo-realistic visualizations of the data collected by such modeling systems. While data acquisition times for mobile mapping systems are considerably shorter than for static ones, their recovered camera poses often suffer from inaccuracies, resulting in visible discontinuities when successive images are projected onto a surface for texturing. We present a method for texture mapping models of indoor environments that starts by selecting images whose camera poses are well-aligned in two dimensions. We then align images to geometry as well as to each other, producing visually consistent textures even in the presence of inaccurate surface geometry and noisy camera poses. Images are then composited into a final texture mosaic and projected onto surface geometry for visualization. The effectiveness of the proposed method is demonstrated on a number of different indoor environments.
© (2014) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Peter Cheng, Michael Anderson, Stewart He, and Avideh Zakhor "Texture mapping 3D models of indoor environments with noisy camera poses", Proc. SPIE 9020, Computational Imaging XII, 90200V (7 March 2014);


View-dependent progressive mesh coding based on partitioning
Proceedings of SPIE (January 04 2002)
Proceedings of SPIE (May 01 2007)
Review of image-based rendering techniques
Proceedings of SPIE (May 30 2000)

Back to Top