Jin, Jing and Xu, Ruitian and Daly, Ian and Zhao, Xueqing and Wang, Xingyu and Cichocki, Andrzej (2024) MOCNN: A Multi-scale Deep Convolutional Neural Network for ERP-based Brain-Computer Interfaces. IEEE Transactions on Cybernetics, 54 (9). pp. 5565-5576. DOI https://doi.org/10.1109/TCYB.2024.3390805
Jin, Jing and Xu, Ruitian and Daly, Ian and Zhao, Xueqing and Wang, Xingyu and Cichocki, Andrzej (2024) MOCNN: A Multi-scale Deep Convolutional Neural Network for ERP-based Brain-Computer Interfaces. IEEE Transactions on Cybernetics, 54 (9). pp. 5565-5576. DOI https://doi.org/10.1109/TCYB.2024.3390805
Jin, Jing and Xu, Ruitian and Daly, Ian and Zhao, Xueqing and Wang, Xingyu and Cichocki, Andrzej (2024) MOCNN: A Multi-scale Deep Convolutional Neural Network for ERP-based Brain-Computer Interfaces. IEEE Transactions on Cybernetics, 54 (9). pp. 5565-5576. DOI https://doi.org/10.1109/TCYB.2024.3390805
Abstract
Event-related potentials (ERPs) reflect neurophysiological changes of the brain in response to external events and their associated underlying complex spatiotemporal feature information is governed by ongoing oscillatory activity within the brain. Deep learning methods have been increasingly adopted for ERP-based brain-computer interfaces (BCIs) due to their excellent feature representation abilities, which allow for deep analysis of oscillatory activity within the brain. Features with higher spatiotemporal frequencies usually represent detailed and localized information, while features with lower spatiotemporal frequencies usually represent global structures. Mining EEG features from multiple spatiotemporal frequencies is conducive to obtaining more discriminative information. A multiscale feature fusion octave convolution neural network (MOCNN) is proposed in this article. MOCNN divides the ERP signals into high-, medium-and low-frequency components corresponding to different resolutions and processes them in different branches. By adding mid-and low-frequency components, the feature information used by MOCNN can be enriched, and the required amount of calculations can be reduced. After successive feature mapping using temporal and spatial convolutions, MOCNN realizes interactive learning among different components through the exchange of feature information among branches. Classification is accomplished by feeding the fused deep spatiotemporal features from various components into a fully connected layer. The results, obtained on two public datasets and a self-collected ERP dataset, show that MOCNN can achieve state-of-the-art ERP classification performance. In this study, the generalized concept of octave convolution is introduced into the field of ERP-BCI research, which allows effective spatiotemporal features to be extracted from multiscale networks through branch width optimization and information interaction at various scales.
Item Type: | Article |
---|---|
Uncontrolled Keywords: | Brain-computer interfaces (BCIs); deep learning; event-related potentials (ERPs); multiscale; octave convolution (OctConv) neural network |
Divisions: | Faculty of Science and Health Faculty of Science and Health > Computer Science and Electronic Engineering, School of |
SWORD Depositor: | Unnamed user with email elements@essex.ac.uk |
Depositing User: | Unnamed user with email elements@essex.ac.uk |
Date Deposited: | 08 Jul 2024 15:23 |
Last Modified: | 30 Oct 2024 21:22 |
URI: | http://repository.essex.ac.uk/id/eprint/38276 |
Available files
Filename: FINAL VERSION.pdf