Please use this identifier to cite or link to this item: http://dspace.iitrpr.ac.in:8080/xmlui/handle/123456789/3653
Title: EEG-ConvTransformer for single-trial EEG-based visual stimulus classification
Authors: Bagchi, S.
Bathula, D.R.
Keywords: Deep learning
EEG
Head representations
Inter-head diversity
Inter-region similarity
Multi-head attention
Temporal convolution
Transformer
Visual stimulus classification
Issue Date: 17-Jul-2022
Abstract: Different categories of visual stimuli evoke distinct activation patterns in the human brain. These patterns can be captured with EEG for utilization in application such as Brain-Computer Interface (BCI). However, accurate classification of these patterns acquired using single-trial data is challenging due to the low signal-to-noise ratio of EEG. Recently, deep learning-based transformer models with multi-head self-attention have shown great potential for analyzing variety of data. This work introduces an EEG-ConvTranformer network that is based on both multi-headed self-attention and temporal convolution. The novel architecture incorporates self-attention modules to capture inter-region interaction patterns and convolutional filters to learn temporal patterns in a single module. Experimental results demonstrate that EEG-ConvTransformer achieves improved classification accuracy over state-of-the-art techniques across five different visual stimulus classification tasks. Finally, quantitative analysis of inter-head diversity also shows low similarity in representational space, emphasizing the implicit diversity of multi-head attention.
URI: http://localhost:8080/xmlui/handle/123456789/3653
Appears in Collections:Year-2022

Files in This Item:
File Description SizeFormat 
Full Text.pdf1.6 MBAdobe PDFView/Open    Request a copy


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.