Discovering Salient Regions on 3D Photo-textured Maps: Crowdsourcing Interaction Data from Multitouch Smartphones and Tablets (bibtex)
by M. Johnson-Roberson, Mitch Bryson, Bertrand Douillard, Oscar Pizarro, Stefan B. Williams
Abstract:
Abstract This paper presents a system for crowdsourcing saliency interest points for 3D photo-textured maps rendered on smartphones and tablets. An app was created that is capable of interactively rendering 3D reconstructions gathered with an Autonomous Underwater Vehicle. Through hundreds of thousands of logged user interactions with the models we attempt to data-mine salient interest points. To this end we propose two models for calculating saliency from human interaction with the data. The first uses the view frustum of the camera to track the amount of time points are on screen. The second uses the velocity of the camera as an indicator of saliency and uses a Hidden Markov model to learn the classification of salient and non-salient points. To provide a comparison to existing techniques several traditional visual saliency approaches are applied to orthographic views of the models' photo-texturing. The results of all approaches are validated with human attention ground truth gathered using a remote gaze-tracking system that recorded the locations of the person's attention while exploring the models.
Reference:
Discovering Salient Regions on 3D Photo-textured Maps: Crowdsourcing Interaction Data from Multitouch Smartphones and Tablets (M. Johnson-Roberson, Mitch Bryson, Bertrand Douillard, Oscar Pizarro, Stefan B. Williams), In Computer Vision and Image Understanding, volume 131, 2015.
Bibtex Entry:
@article{Johnson-Roberson:2015aa,
	Abstract = {Abstract This paper presents a system for crowdsourcing saliency interest points for 3D photo-textured maps rendered on smartphones and tablets. An app was created that is capable of interactively rendering 3D reconstructions gathered with an Autonomous Underwater Vehicle. Through hundreds of thousands of logged user interactions with the models we attempt to data-mine salient interest points. To this end we propose two models for calculating saliency from human interaction with the data. The first uses the view frustum of the camera to track the amount of time points are on screen. The second uses the velocity of the camera as an indicator of saliency and uses a Hidden Markov model to learn the classification of salient and non-salient points. To provide a comparison to existing techniques several traditional visual saliency approaches are applied to orthographic views of the models' photo-texturing. The results of all approaches are validated with human attention ground truth gathered using a remote gaze-tracking system that recorded the locations of the person's attention while exploring the models.},
	Author = {Johnson-Roberson, M. and Bryson, Mitch and Douillard, Bertrand and Pizarro, Oscar and Williams, Stefan B.},
	Date-Added = {2014-12-12 16:02:56 +0000},
	Date-Modified = {2015-11-29 16:52:35 +0000},
	Journal = {Computer Vision and Image Understanding},
	Keywords = {jrnl},
	Number = {0},
	Pages = {28--41},
	Title = {Discovering Salient Regions on 3D Photo-textured Maps: Crowdsourcing Interaction Data from Multitouch Smartphones and Tablets},
	Volume = {131},
	Year = {2015},
	Bdsk-Url-1 = {http://www.sciencedirect.com/science/article/pii/S1077314214001593},
	Bdsk-Url-2 = {http://dx.doi.org/10.1016/j.cviu.2014.07.006}}
Powered by bibtexbrowser