Skip to content

eeyhsong/EEG-Conformer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

40 Commits
 
 
 
 
 
 
 
 

Repository files navigation

EEG-Conformer

EEG Conformer: Convolutional Transformer for EEG Decoding and Visualization [Paper]

Core idea: spatial-temporal conv + pooling + self-attention

News

🎉🎉🎉 We've joined in braindecode toolbox. Use here for detailed info.

Thanks to Bru and colleagues for helping with the modifications.

Abstract

Network Architecture

  • We propose a compact convolutional Transformer, EEG Conformer, to encapsulate local and global features in a unified EEG classification framework.
  • The convolution module learns the low-level local features throughout the one-dimensional temporal and spatial convolution layers. The self-attention module is straightforwardly connected to extract the global correlation within the local temporal features. Subsequently, the simple classifier module based on fully-connected layers is followed to predict the categories for EEG signals.
  • We also devise a visualization strategy to project the class activation mapping onto the brain topography.

Requirements:

  • Python 3.10
  • Pytorch 1.12

Datasets

Please use consistent train-val-test split when comparing with other methods.

Citation

Hope this code can be useful. I would appreciate you citing us in your paper. 😊

@article{song2023eeg,
  title = {{{EEG Conformer}}: {{Convolutional Transformer}} for {{EEG Decoding}} and {{Visualization}}},
  shorttitle = {{{EEG Conformer}}},
  author = {Song, Yonghao and Zheng, Qingqing and Liu, Bingchuan and Gao, Xiaorong},
  year = {2023},
  journal = {IEEE Transactions on Neural Systems and Rehabilitation Engineering},
  volume = {31},
  pages = {710--719},
  issn = {1558-0210},
  doi = {10.1109/TNSRE.2022.3230250}
}

About

EEG Transformer 2.0. i. Convolutional Transformer for EEG Decoding. ii. Novel visualization - Class Activation Topography.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages