Visual dictionaries in the Brain: Comparing HMAX and BOW

Open Access
Authors
Publication date 2014
Book title 2014 IEEE International Conference on Multimedia and Expo (ICME 2014): Chengdu, China 14-18 July 2014
ISBN
  • 9781479947607
Event IEEE International Conference on Multimedia & Expo
Publisher Piscataway, NJ: IEEE
Organisations
  • Faculty of Science (FNWI) - Informatics Institute (IVI)
Abstract
The human visual system is thought to use features of intermediate complexity for scene representation. How the brain computationally represents intermediate features is, however, still unclear. Here we tested and compared two widely used computational models - the biologically plausible HMAX model and Bag of Words (BoW) model from computer vision against human brain activity. These computational models use visual dictionaries, candidate features of intermediate complexity, to represent visual scenes, and the models have been proven effective in automatic object and scene recognition. We analyzed where in the brain and to what extent human fMRI responses to natural scenes can be accounted for by the HMAX and BoW representations. Voxel-wise application of a distance-based variation partitioning method reveals that HMAX explains significant brain activity in early visual regions and also in higher regions such as LO, TO while the BoW primarily explains brain acitvity in the early visual area. Notably, both HMAX and BoW explain the most brain activity in higher areas such as V4 and TO. These results suggest that visual dictionaries might provide a suitable computation for the representation of intermediate features in the brain.
Document type Conference contribution
Language English
Published at https://doi.org/10.1109/ICME.2014.6890312
Downloads
Visual dictionaries in the Brain (Final published version)
Permalink to this page
Back