Relate Auditory Speech to EEG by Shallow-Deep Attention-based Network
Fan Cui (Mi); Liyong Guo (Xiaomi Corp.); Lang He (XUPT); Jiyao Liu (NWPU); Ercheng Pei (XUPT); Yujun Wang (Mi); Dongmei Jiang (Northwestern Polytechnical University \ Peng Cheng Laboratory)
-
SPS
IEEE Members: $11.00
Non-members: $15.00
Electroencephalography (EEG) plays a vital role in detecting how brain responses to different stimulus. In this paper, we propose a novel Shallow-Deep Attention-based Network (SDANet) to classify the correct auditory stimulus evoking the EEG signal. It adopts the Attention-based Correlation Module (ACM) to discover the connection between auditory speech and EEG from global aspect, and the Shallow-Deep Similarity Classification Module (SDSCM) to decide the classification result via the embeddings learned from the shallow and deep layers.
Moreover, various training strategies and data augmentation are used to boost the model robustness. Experiments are conducted on the dataset provided by Auditory EEG challenge (ICASSP Signal Processing Grand Challenge 2023). Results show that the proposed model has a significant gain over the baseline on the match-mismatch track.