High frequency visual stimuli induced functional networks of emotion recognition: an MEG study

Poster No:

1944 

Submission Type:

Abstract Submission 

Authors:

RUOCHU XIONG1, Masashi Kinoshita2, Kenji Yoshiki2, Riho Nakajima3, Mitsutoshi Nakada2

Institutions:

1Department of Neurosurgery, Graduate School of Medical Sciences, Kanazawa University, Kanazawa, Japan, 2Department of Neurosurgery, Kanazawa University, Kanazawa, Japan, 3Department of Occupational Therapy, Kanazawa University, Kanazawa, Japan

First Author:

RUOCHU XIONG  
Department of Neurosurgery, Graduate School of Medical Sciences, Kanazawa University
Kanazawa, Japan

Co-Author(s):

Masashi Kinoshita  
Department of Neurosurgery, Kanazawa University
Kanazawa, Japan
Kenji Yoshiki  
Department of Neurosurgery, Kanazawa University
Kanazawa, Japan
Riho Nakajima  
Department of Occupational Therapy, Kanazawa University
Kanazawa, Japan
Mitsutoshi Nakada  
Department of Neurosurgery, Kanazawa University
Kanazawa, Japan

Introduction:

Researchers have studied emotion recognition with various modalities, but results vary depending on different tasks and analysis methods. In the field of EEG, fast periodic visual stimulation (FPVS) was established to stably activate the functional networks apart from the primary visual cortex (V1) (Rossion & Boremanse, 2011)(Keil et al., 2011). Although MEG studies have proved the validity of FPVS (Minarik et al., 2023)(Parkkonen et al., 2008), most studies chose low frequencies that might entrain endogenous activity. However, high frequency stimuli avoid entrainment of endogenous gamma activity (Duecker et al., 2021). In this study, we applied high-frequency FPVS in MEG for emotion recognition. We hypothesized that high frequency visual stimuli would activate a related functional network, besides V1, during the emotion recognition task.

Methods:

We collected MEG data from 22 healthy subjects. The task started with a 2s emotional word, followed by 2 possible choices of eyes with basic emotions lasting for 4s, then a 2s blank screen as baseline. Subjects were asked to fixate on the figure corresponding to the emotional word. By modulating the amplitude of RGB code, we tagged the figure corresponding to the word with 80Hz, and the inconsistent one with 90Hz. After MEG scan, participants recorded their figure choices on paper.
Using Brainstorm and Matlab, we calculated the averaged time-frequency (TF) map for V1, in the Desikan-Killiany atlas, called "pericalcarine". Source activity at 80Hz, 90Hz, and all frequencies was estimated. White matter tracts were reconstructed based on imaginary coherence results of connectivity analysis. Statistical analysis was t-test comparing the activity of viewing figures and viewing blank (baseline: -4.0000s to -2.0005s), with a significance threshold of 0.0001.
Supporting Image: 1.jpg
 

Results:

Behavioral data confirmed that all subjects correctly recognized the emotion. There was a strong and stable activation at 80Hz from 0s to 4s (viewing figures) on the TF map of V1. On the 80Hz source maps, we divided the activated areas into 5 groups according to the begin of their activation time: 0s~0.1s: bilateral pericalcarine, lateral occipital, cuneus, lingual area; 0.1s~0.2s: bilateral parahippocampal, fusiform, precuneus, and entorhinal area; 0.2s~0.6s: bilateral isthmus cingulate, inferior temporal, inferior parietal, and superior parietal, temporal pole, insular, superior temporal sulcus, posterior cingulate cortex, and middle temporal gyrus; 0.6s~1.0s: right parstriangularis, and left medial orbitofrontal area; 1.0s~: right superior temporal, bilateral paracentral, right lateral orbitofrontal, right postcentral, right precentral, and right caudal anterior cingulate area. From Group 1 to 5, the maximum statistic value of these activation areas gradually decreased. These groups roughly corresponded to five stages of task performance: visual processing, face recognition and memory retrieving, emotion perceiving, execution of selection task, and maintenance of attention. While no clear 90Hz activity was observed in the TF map, the 90Hz source map indicated activations similar to, but weaker than those at 80Hz. The source map of all frequencies showed that the activity was mainly in 0.6s from 0s. Connectivity analysis highlighted the role of the corpus callosum and anterior commissure in facial emotion recognition.
Supporting Image: Fig3andFig4.jpg
 

Conclusions:

In this study, we identified a bilateral temporospatial network for facial emotion recognition, mediated by the splenium of corpus callosum and the anterior commissure. This network spanned the occipital lobes, temporal lobes, cingulate gyrus, and insula-likely core regions for processing facial expressions-before engaging the frontal lobe to execute the task. Our findings suggest that the upper limit of usable tagging frequency may depend on task complexity; thus, researchers should optimize the frequency for their specific design.

Emotion, Motivation and Social Neuroscience:

Emotional Perception 2

Novel Imaging Acquisition Methods:

MEG 1

Keywords:

Emotions
MEG
Other - fast periodic visual stimulation

1|2Indicates the priority used for review

Abstract Information

By submitting your proposal, you grant permission for the Organization for Human Brain Mapping (OHBM) to distribute your work in any format, including video, audio print and electronic text through OHBM OnDemand, social media channels, the OHBM website, or other electronic publications and media.

I accept

The Open Science Special Interest Group (OSSIG) is introducing a reproducibility challenge for OHBM 2025. This new initiative aims to enhance the reproducibility of scientific results and foster collaborations between labs. Teams will consist of a “source” party and a “reproducing” party, and will be evaluated on the success of their replication, the openness of the source work, and additional deliverables. Click here for more information. Propose your OHBM abstract(s) as source work for future OHBM meetings by selecting one of the following options:

I do not want to participate in the reproducibility challenge.

Please indicate below if your study was a "resting state" or "task-activation” study.

Task-activation

Healthy subjects only or patients (note that patient studies may also involve healthy subjects):

Healthy subjects

Was this research conducted in the United States?

No

Were any human subjects research approved by the relevant Institutional Review Board or ethics panel? NOTE: Any human subjects studies without IRB approval will be automatically rejected.

Yes

Were any animal research approved by the relevant IACUC or other animal research panel? NOTE: Any animal studies without IACUC approval will be automatically rejected.

Not applicable

Please indicate which methods were used in your research:

MEG

For human MRI, what field strength scanner do you use?

1.5T

Which processing packages did you use for your study?

Other, Please list  -   Brainstorm

Provide references using APA citation style.

1. Duecker, K., Gutteling, T. P., Herrmann, C. S., & Jensen, O. (2021). No Evidence for Entrainment: Endogenous Gamma Oscillations and Rhythmic Flicker Responses Coexist in Visual Cortex. The Journal of Neuroscience, 41(31), 6684–6698. https://doi.org/10.1523/JNEUROSCI.3134-20.2021
2. Keil, A., Costa, V., Smith, J. C., Sabatinelli, D., McGinnis, E. M., Bradley, M. M., & Lang, P. J. (2011). Tagging cortical networks in emotion: A topographical analysis. Human Brain Mapping, 33(12), 2920–2931. https://doi.org/10.1002/hbm.21413
3. Minarik, T., Berger, B., & Jensen, O. (2023). Optimal parameters for rapid (invisible) frequency tagging using MEG. NeuroImage, 281, 120389. https://doi.org/10.1016/j.neuroimage.2023.120389
4. Parkkonen, L., Andersson, J., Hämäläinen, M., & Hari, R. (2008). Early visual brain areas reflect the percept of an ambiguous scene. Proceedings of the National Academy of Sciences, 105(51), 20500–20504. https://doi.org/10.1073/pnas.0810966105
5. Rossion, B., & Boremanse, A. (2011). Robust sensitivity to facial identity in the right human occipito-temporal cortex as revealed by steady-state visual-evoked potentials. Journal of Vision, 11(2), 16–16. https://doi.org/10.1167/11.2.16

UNESCO Institute of Statistics and World Bank Waiver Form

I attest that I currently live, work, or study in a country on the UNESCO Institute of Statistics and World Bank List of Low and Middle Income Countries list provided.

No