We present the results of a first experimental study to improve the computation of saliency maps, by using luminance and depth images features. More specifically, we have recorded the center of gaze of users when they were viewing natural scenes. We used machine learning techniques to train a bottom-up, top-down model of saliency based on 2D and depth features/cues. We found that models trained on Itti & Koch and depth features combined outperform models trained on other individual features (i.e. only Gabor filter responses or only depth features), or trained on combination of these features. As a consequence, depth features combined with Itti & Koch features improve the prediction of gaze locations. This first characterization of using joint luminance and depth features is an important step towards developing models of eye movements, which operate well under natural conditions such as those encountered in HCI settings. © 2014 Springer International Publishing.
CITATION STYLE
Mohammed, R. A. A., Schwabe, L., & Staadt, O. (2014). Gaze location prediction with depth features as auxiliary information. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8511 LNCS, pp. 281–292). Springer Verlag. https://doi.org/10.1007/978-3-319-07230-2_28
Mendeley helps you to discover research relevant for your work.