Papers
Topics
Authors
Recent
2000 character limit reached

EEG-Transformer: Self-attention from Transformer Architecture for Decoding EEG of Imagined Speech

Published 15 Dec 2021 in cs.HC, cs.SD, and eess.AS | (2112.09239v1)

Abstract: Transformers are groundbreaking architectures that have changed a flow of deep learning, and many high-performance models are developing based on transformer architectures. Transformers implemented only with attention with encoder-decoder structure following seq2seq without using RNN, but had better performance than RNN. Herein, we investigate the decoding technique for electroencephalography (EEG) composed of self-attention module from transformer architecture during imagined speech and overt speech. We performed classification of nine subjects using convolutional neural network based on EEGNet that captures temporal-spectral-spatial features from EEG of imagined speech and overt speech. Furthermore, we applied the self-attention module to decoding EEG to improve the performance and lower the number of parameters. Our results demonstrate the possibility of decoding brain activities of imagined speech and overt speech using attention modules. Also, only single channel EEG or ear-EEG can be used to decode the imagined speech for practical BCIs.

Citations (37)

Summary

  • The paper presents a hybrid EEG-Transformer model that integrates convolutional features with self-attention to decode imagined and overt speech EEG signals.
  • It shows an average accuracy of 49.5% for overt speech and 35.07% for imagined speech, demonstrating significant improvements over traditional methods.
  • The study highlights the potential for lightweight, portable BCI systems by efficiently processing limited-channel EEG signals using Transformer architecture.

EEG-Transformer: Self-Attention for Decoding Imagined Speech EEG

The paper "EEG-Transformer: Self-attention from Transformer Architecture for Decoding EEG of Imagined Speech" (2112.09239) presents an innovative application of Transformer architectures in the context of EEG signal decoding for imagined and overt speech tasks. Leveraging the self-attention mechanism, the authors aim to enhance classification accuracy while reducing model complexity, showcasing the viability of attention-based models for brain-computer interface (BCI) applications.

Introduction to EEG Decoding and Transformers

The advent of deep learning has transformed traditional approaches to decoding electroencephalography (EEG) signals, particularly in brain-computer interfaces (BCIs) tasked with recognizing cognitive states or intentions. This paper explores the potential of Transformer architectures, specifically applying self-attention modules to the temporal-spectral-spatial data derived from EEG. Unlike recurrent neural networks (RNNs), Transformers facilitate parallel processing, addressing limitations in computation speed and enabling robust, interpretable models for EEG signal analysis. Figure 1

Figure 1: Total frameworks in this study. We split an image into fixed-size patches, linearly embed each of them, add position embeddings, and feed the resulting sequence of vectors to a standard Transformer encoder.

Data and Preprocessing

The research involved nine participants, with EEG data collected during tasks of imagined and overt speech involving 12 distinct words and a resting state. The preprocessing pipeline included down-sampling EEG signals, applying a high-gamma band Butterworth filter, and executing baseline correction. Channels around the Broca and Wernicke areas were specifically targeted, and independent component analysis mitigated artifacts from electromyography (EMG) and ocular signals.

Architecture Overview

The EEG-Transformer framework integrates convolutional layers for extracting features and incorporates self-attention mechanisms to exploit long-range dependencies within the EEG data. This hybrid model architecture not only compresses model size but also enhances classification accuracy through multi-head attention, which resolves EEG representations into distinct subspaces. Figure 2

Figure 2: Transformer architecture. Self-attention and feed-forward networks with subsystem integrations, optimizing long-term dependency path lengths.

Results and Performance Evaluation

The Transformer-based model exhibited noteworthy performance, with overt speech reaching an average accuracy of 49.5% across 13 classes, while imagined speech scored 35.07%. Although overt speech naturally encompasses richer features due to associative EMG activities, the marginal performance gap indicates the potential of imagined speech decoding. Statistically significant improvement was noted with the self-attention module, highlighting its effectiveness in focusing computational resources on relevant EEG features.

Discussion on Practical Implications

This research underscores the feasibility of employing Transformers in EEG-based imagined speech decoding, with critical implications for real-world BCIs. Notably, the ability to achieve high performance with limited channels, such as ear-EEG configurations, proposes a robust, practical communication system applicable to portable BCI technologies. Future work may explore optimizing architectural parameters and expanding datasets to refine the accuracy and generalization capabilities of Transformer models in neural signal processing.

Conclusion

The study successfully applies self-attention from Transformer models to EEG decoding tasks, demonstrating significant advancements in classification performance for imagined speech. These findings provide a compelling case for integrating self-attention mechanisms into practical BCI frameworks, promoting the development of more responsive and accurate neuro-interfacing systems. Future research directions could involve investigating other Transformer-derived mechanisms, such as Vision Transformers, to further augment the efficiency and versatility of EEG decoding strategies in varied BCI applications.

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Authors (2)

Collections

Sign up for free to add this paper to one or more collections.