TY - GEN
T1 - The semantic paintbrush
T2 - 33rd Annual CHI Conference on Human Factors in Computing Systems, CHI 2015
AU - Miksik, Ondrej
AU - Vineet, Vibhav
AU - Lidegaard, Morten
AU - Prasaath, Ram
AU - Nießner, Matthias
AU - Golodetz, Stuart
AU - Hicks, Stephen L.
AU - Pérez, Patrick
AU - Izadi, Shahram
AU - Torr, Philip H.S.
N1 - Publisher Copyright:
© Copyright 2015 ACM.
PY - 2015/4/18
Y1 - 2015/4/18
N2 - We present an augmented reality system for large scale 3D reconstruction and recognition in outdoor scenes. Unlike existing prior work, which tries to reconstruct scenes using active depth cameras, we use a purely passive stereo setup, allowing for outdoor use and extended sensing range. Our system not only produces a map of the 3D environment in real-time, it also allows the user to draw (or 'paint') with a laser pointer directly onto the reconstruction to segment the model into objects. Given these examples our system then learns to segment other parts of the 3D map during online acquisition. Unlike typical object recognition systems, ours therefore very much places the user 'in the loop' to segment particular objects of interest, rather than learning from predefined databases. The laser pointer additionally helps to 'clean up' the stereo reconstruction and final 3D map, interactively. Using our system, within minutes, a user can capture a full 3D map, segment it into objects of interest, and refine parts of the model during capture. We provide full technical details of our system to aid replication, as well as quantitative evaluation of system components. We demonstrate the possibility of using our system for helping the visually impaired navigate through spaces. Beyond this use, our system can be used for playing large-scale augmented reality games, shared online to augment streetview data, and used for more detailed car and person navigation.
AB - We present an augmented reality system for large scale 3D reconstruction and recognition in outdoor scenes. Unlike existing prior work, which tries to reconstruct scenes using active depth cameras, we use a purely passive stereo setup, allowing for outdoor use and extended sensing range. Our system not only produces a map of the 3D environment in real-time, it also allows the user to draw (or 'paint') with a laser pointer directly onto the reconstruction to segment the model into objects. Given these examples our system then learns to segment other parts of the 3D map during online acquisition. Unlike typical object recognition systems, ours therefore very much places the user 'in the loop' to segment particular objects of interest, rather than learning from predefined databases. The laser pointer additionally helps to 'clean up' the stereo reconstruction and final 3D map, interactively. Using our system, within minutes, a user can capture a full 3D map, segment it into objects of interest, and refine parts of the model during capture. We provide full technical details of our system to aid replication, as well as quantitative evaluation of system components. We demonstrate the possibility of using our system for helping the visually impaired navigate through spaces. Beyond this use, our system can be used for playing large-scale augmented reality games, shared online to augment streetview data, and used for more detailed car and person navigation.
KW - 3D reconstruction
KW - Augmented reality
KW - Laser pointer interaction
KW - Semantic segmentation
KW - Stereo
KW - Visually impaired
UR - http://www.scopus.com/inward/record.url?scp=84951059759&partnerID=8YFLogxK
U2 - 10.1145/2702123.2702222
DO - 10.1145/2702123.2702222
M3 - Conference contribution
AN - SCOPUS:84951059759
T3 - Conference on Human Factors in Computing Systems - Proceedings
SP - 3317
EP - 3326
BT - CHI 2015 - Proceedings of the 33rd Annual CHI Conference on Human Factors in Computing Systems
PB - Association for Computing Machinery
Y2 - 18 April 2015 through 23 April 2015
ER -