Humans can attend to and categorise objects individually, but also as groups. We present a computational model of how visual attention is allocated to single objects and groups of objects, and how single objects and groups are classified. We illustrate the model with a novel account of the role of stimulus similarity in visual search tasks, as identified by Duncan and Humphreys [1]. © Springer International Publishing 2013.
CITATION STYLE
Walles, H., Robins, A., & Knott, A. (2013). A neural network model of visual attention and group classification, and its performance in a visual search task. In Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) (Vol. 8272 LNAI, pp. 98–103). https://doi.org/10.1007/978-3-319-03680-9_11
Mendeley helps you to discover research relevant for your work.