4.3 Article

Comparing visual representations across human fMRI and computational vision

Journal

JOURNAL OF VISION
Volume 13, Issue 13, Pages -

Publisher

ASSOC RESEARCH VISION OPHTHALMOLOGY INC
DOI: 10.1167/13.13.25

Keywords

neuroimaging; object recognition; computational modeling; intermediate feature representation

Categories

Funding

  1. NIH EUREKA [1R01MH084195-01]
  2. Temporal Dynamic of Learning Center at UCSD (NSF Science of Learning Center) [SBE-0542013]
  3. NSF IGERT
  4. R. K. Mellon Foundation
  5. Program in Neural Computation (NIH) [T90 DA022762]
  6. Pennsylvania Department of Health's Commonwealth Universal Research Enhancement Program

Ask authors/readers for more resources

Feedforward visual object perception recruits a cortical network that is assumed to be hierarchical, progressing from basic visual features to complete object representations. However, the nature of the intermediate features related to this transformation remains poorly understood. Here, we explore how well different computer vision recognition models account for neural object encoding across the human cortical visual pathway as measured using fMRI. These neural data, collected during the viewing of 60 images of real-world objects, were analyzed with a searchlight procedure as in Kriegeskorte, Goebel, and Bandettini (2006): Within each searchlight sphere, the obtained patterns of neural activity for all 60 objects were compared to model responses for each computer recognition algorithm using representational dissimilarity analysis (Kriegeskorte et al., 2008). Although each of the computer vision methods significantly accounted for some of the neural data, among the different models, the scale invariant feature transform (Lowe, 2004), encoding local visual properties gathered from interest points, was best able to accurately and consistently account for stimulus representations within the ventral pathway. More generally, when present, significance was observed in regions of the ventral-temporal cortex associated with intermediate-level object perception. Differences in model effectiveness and the neural location of significant matches may be attributable to the fact that each model implements a different featural basis for representing objects (e. g., more holistic or more parts-based). Overall, we conclude that well-known computer vision recognition systems may serve as viable proxies for theories of intermediate visual object representation.

Authors

I am an author on this paper
Click your name to claim this paper and add it to your profile.

Reviews

Primary Rating

4.3
Not enough ratings

Secondary Ratings

Novelty
-
Significance
-
Scientific rigor
-
Rate this paper

Recommended

No Data Available
No Data Available