Presented During:
Wednesday, June 26, 2024: 11:30 AM - 12:45 PM
COEX
Room:
ASEM Ballroom 202
Poster No:
200
Submission Type:
Abstract Submission
Authors:
Chris Kang1, Jasmine Moore1, Matthias Wilms1, Nils Forkert1
Institutions:
1University of Calgary, Calgary, Alberta
First Author:
Co-Author(s):
Introduction:
Convolutional neural networks (CNNs) have emerged as a popular choice of deep learning architecture for modeling visual processing, as their hierarchical structure and flow of information processing closely resembles the human ventral stream [LeCun, Y. (1989)]. While CNNs have been used to model healthy visual cognition, there remain limitations in biologically plausible in silico modeling of cognitive decline in neurodegenerative diseases, such as Alzheimer's (AD). Previously, we developed methods to simulate neurodegeneration of the visual system through iterative synaptic injury in CNNs [Tuladhar, A. (2021), Moore, J. (2023)]. However, the limitation of CNNs lies in the lack of biologically meaningful learning mechanisms that are similar to cognitive functions, such as memory. These mechanisms are essential for accurately capturing the neuropathogenesis. For instance, the deposition of beta-amyloid peptide and neurofibrillary tangles of tau polymers in the hippocampus leads to cognitive decline in memory tasks among AD patients.
Building on our prior work, in this study, we equipped a CNN with associative memory to enhance biological plausibility, combining two critical cognitive functions of the brain: visual processing of the ventral stream and associative memory of the hippocampus. The model demonstrates intriguing and beneficial properties, including (1) robustness to noisy or occluded image queries and (2) interpretable and sparse representations in network weights. We argue that this model is an improved in silico framework for a healthy brain, as well as the cognitive profiles of AD progression.
Methods:
VGG19, a CNN with a high similarity to the human brain as measured by Brain-Score [Schrimpf, M. (2018)], was equipped with a flattened layer of the modern Hopfield network [Krotov, D. (2016), Ramsauer, H. (2020)], replacing the penultimate fully-connected layer (Figure 1a). This VGG-MHN model was independently trained on two commonly used vision datasets, MNIST and CIFAR-10, for image classification.
Previously trained CIFAR-10 images and test images were injected with Gaussian noise, according to varying signal-to-signal plus noise ratios, and classification tasks were performed to measure the ability to recall noisy queries, a known ability of the human brain. The result was compared to the baseline performance of VGG19, which has been shown to perform poorly when inputs are even slightly altered [Tang, H. (2018)]. For a more challenging recall task, training and classification with MNIST images was performed on partially occluded test sets featuring images masked by 30% and 50% of the total area, and the network weights were analyzed.
Results:
VGG-MHN exhibited significantly improved robustness to noisy queries for previously seen images compared to the baseline VGG19 (Figure 1b). In case of unseen images, both VGG-MHN and VGG19 performed equally in the high SSNR domain. However, VGG-MHN underperformed when the SSNR level was below w=0.8. In the occluded MNIST experiment, VGG-MHN significantly outperformed VGG19 (Figure 1c).
Further analysis of the VGG-MHN when tested on the occluded MNIST dataset revealed that, in the Hopfield layer, the weights preserve feature and prototype representations depending on the model choice of energy function. This phenomenon was first characterized as a "feature-to-prototype" transition (Figure 1d) [Krotov, D. (2018)].
Conclusions:
The biologically inspired CNN, equipped with associative memory, extends our existing framework for in silico neurodegeneration. VGG-HMN effectively integrates visual processing with memory systems grounded on fundamental cognitive principles (i.e., Hebbian learning) and offers many advantages, simulating a healthy brain. These include enhanced robustness against noisy and occluded queries and the production of interpretable representations. We believe that such an architecture is well suited for in silico analysis of neurodegenerative diseases in the forthcoming work.
Disorders of the Nervous System:
Neurodegenerative/ Late Life (eg. Parkinson’s, Alzheimer’s) 1
Learning and Memory:
Neural Plasticity and Recovery of Function
Learning and Memory Other 2
Neuroinformatics and Data Sharing:
Informatics Other
Perception, Attention and Motor Behavior:
Attention: Visual
Keywords:
Machine Learning
Memory
STRUCTURAL MRI
1|2Indicates the priority used for review
Provide references using author date format
Krotov, D. (2016). 'Dense associative memory for pattern recognition', Advances in Neural Information Processing Systems, 29
Krotov, D. (2018). 'Dense associative memory is robust to adversarial inputs', Neural Computation, 30(12):3151–3167
LeCun, Y. (1989). 'Backpropagation applied to handwritten zip code recognition', Neural Computation, 1(4):541–551
Moore, J. A. (2023a). 'Dementia in convolutional neural networks: Using deep learning models to simulate neurodegeneration of the visual system', Neuroinformatics, 21(1):45–55
Moore, J. A. (2023b). 'Simulation of neuroplasticity in a CNN-based in-silico model of neurodegeneration of the visual system', Frontiers in Computational Neuroscience, 17
Ramsauer, H. (2020). 'Hopfield networks is all you need', arXiv preprint arXiv:2008.02217
Schrimpf, M. (2018). 'Brain-score: Which artificial neural network for object recognition is most brain-like?', BioRxiv, page 407007
Tang, H. (2018). 'Recurrent computations for visual pattern completion', Proceedings of the National Academy of Sciences, 115(35):8835–8840
Tuladhar, A. (2021). 'Modeling neurodegeneration in silico with deep learning', Frontiers in Neuroinformatics, 15:748370