4.6 Article

Where to Look Next? Combining Static and Dynamic Proto-objects in a TVA-based Model of Visual Attention

Journal

COGNITIVE COMPUTATION
Volume 2, Issue 4, Pages 326-343

Publisher

SPRINGER
DOI: 10.1007/s12559-010-9080-1

Keywords

Modeling visual attention; TVA; Proto-objects; Static and dynamic features; Inhomogeneity; Natural scenes; Top-down control

Funding

  1. Cluster of Excellence - Cognitive Interaction Technology (CITEC)

Ask authors/readers for more resources

To decide Where to look next ? is a central function of the attention system of humans, animals and robots. Control of attention depends on three factors, that is, low-level static and dynamic visual features of the environment (bottom-up), medium-level visual features of proto-objects and the task (top-down). We present a novel integrated computational model that includes all these factors in a coherent architecture based on findings and constraints from the primate visual system. The model combines spatially inhomogeneous processing of static features, spatio-temporal motion features and task-dependent priority control in the form of the first computational implementation of saliency computation as specified by the Theory of Visual Attention (TVA, [7]). Importantly, static and dynamic processing streams are fused at the level of visual proto-objects, that is, ellipsoidal visual units that have the additional medium-level features of position, size, shape and orientation of the principal axis. Proto-objects serve as input to the TVA process that combines top-down and bottom-up information for computing attentional priorities so that relatively complex search tasks can be implemented. To this end, separately computed static and dynamic proto-objects are filtered and subsequently merged into one combined map of proto-objects. For each proto-object, attentional priorities in the form of attentional weights are computed according to TVA. The target of the next saccade is the center of gravity of the proto-object with the highest weight according to the task. We illustrate the approach by applying it to several real world image sequences and show that it is robust to parameter variations.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.6
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available