Neural representations of the perception of handwritten digits and visual objects from a convolutional neural network compared to humans

We investigated neural representations for visual perception of 10 handwritten digits and six visual objects from a convolutional neural network (CNN) and humans using functional magnetic resonance imaging (fMRI). Once our CNN model was fine‐tuned using a pre‐trained VGG16 model to recognize the vis...

Full description

Saved in:
Bibliographic Details
Published in:Human brain mapping Vol. 44; no. 5; pp. 2018 - 2038
Main Authors: Lee, Juhyeon, Jung, Minyoung, Lustig, Niv, Lee, Jong‐Hwan
Format: Journal Article
Language:English
Published: Hoboken, USA John Wiley & Sons, Inc 01-04-2023
Subjects:
Online Access:Get full text
Tags: Add Tag
No Tags, Be the first to tag this record!
Description
Summary:We investigated neural representations for visual perception of 10 handwritten digits and six visual objects from a convolutional neural network (CNN) and humans using functional magnetic resonance imaging (fMRI). Once our CNN model was fine‐tuned using a pre‐trained VGG16 model to recognize the visual stimuli from the digit and object categories, representational similarity analysis (RSA) was conducted using neural activations from fMRI and feature representations from the CNN model across all 16 classes. The encoded neural representation of the CNN model exhibited the hierarchical topography mapping of the human visual system. The feature representations in the lower convolutional (Conv) layers showed greater similarity with the neural representations in the early visual areas and parietal cortices, including the posterior cingulate cortex. The feature representations in the higher Conv layers were encoded in the higher‐order visual areas, including the ventral/medial/dorsal stream and middle temporal complex. The neural representations in the classification layers were observed mainly in the ventral stream visual cortex (including the inferior temporal cortex), superior parietal cortex, and prefrontal cortex. There was a surprising similarity between the neural representations from the CNN model and the neural representations for human visual perception in the context of the perception of digits versus objects, particularly in the primary visual and associated areas. This study also illustrates the uniqueness of human visual perception. Unlike the CNN model, the neural representation of digits and objects for humans is more widely distributed across the whole brain, including the frontal and temporal areas. We investigated neural representations for visual perception of 10 handwritten digits and six natural objects using representational similarity analysis between features from a convolutional neural network (CNN) and neuronal activations from functional magnetic resonance imaging. There was a surprising similarity between the neural representations from the CNN model and neural representations for human visual perception, particularly in the primary visual and associated areas. Moreover, unlike CNN, the neural representation of digits and objects for humans is more widely distributed across the whole brain, including the frontal and temporal areas.
Bibliography:Funding information
Electronics and Telecommunications Research Institute, Grant/Award Number: 22ZS1100; National Research Foundation of Korea, Grant/Award Numbers: 2017R1E1A1A01077288, 2021M3E5D2A01022515
Juhyeon Lee and Minyoung Jung contributed equally.
ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 23
Funding information Electronics and Telecommunications Research Institute, Grant/Award Number: 22ZS1100; National Research Foundation of Korea, Grant/Award Numbers: 2017R1E1A1A01077288, 2021M3E5D2A01022515
ISSN:1065-9471
1097-0193
DOI:10.1002/hbm.26189