- Visual word representation in the brain
- 3rd NIPS Workshop on Machine Learning and Interpretation in NeuroImaging (MLINI '13)
- Book/source title
- MLINI '13: machine learning and interpretation in neuroimaging: proceedings of MLINI 2013
- Number of pages
- Document type
- Conference contribution
- Faculty of Social and Behavioural Sciences (FMG)
Faculty of Science (FNWI)
- Psychology Research Institute (PsyRes)
Informatics Institute (IVI)
The human visual system is thought to use features of intermediate complexity for scene representation. How the brain computationally represents intermediate features is unclear, however. To study this, we tested the Bag of Words (BoW) model in computer vision against human brain activity. This computational model uses visual word histograms, candidate features of intermediate complexity, to represent visual scenes, and has proven effective in automatic object and scene recognition. We analyzed where in the brain and to what extent human fMRI responses to natural scenes can be accounted for by BoW representations. Voxel-wise application of a distance-based variation partitioning method reveals that BoW representations explain brain activity in visual areas V1, V2 and in particular V4. Area V4 is known to be tuned for features of intermediate complexity, suggesting that the BoW model captures intermediate-level scene representations in the human brain.
- Final publisher version
- Other links
If you believe that digital publication of certain material infringes any of your rights or (privacy) interests, please let the Library know, stating your reasons. In case of a legitimate complaint, the Library will make the material inaccessible and/or remove it from the website. Please Ask the Library, or send a letter to: Library of the University of Amsterdam, Secretariat, Singel 425, 1012 WP Amsterdam, The Netherlands. You will be contacted as soon as possible.