Pytorch codes of our paper "Improving Cross-State and Cross-Subject Visual ERP-based BCI with Temporal Modeling and Adversarial Training", which is published in TNSRE.
Brain-computer interface (BCI) is a useful device for people without relying on peripheral nerves and muscles. However, the performance of the event-related potential (ERP)-based BCI declines when applying it to the real environments, especially in cross-state and cross-subject conditions.
For addressing this problem, we propose a general method with hierarchical temporal modeling and adversarial training to enhance the performance of the visual ERP-based BCI under different auditory workload states. The artitecture of our model is shown below.
The rationality of our method here is that the ERP-based BCI is based on electroencephalography (EEG) signals recorded from the surface of the scalp, continuously changing with time and somewhat stochastic. Therefore, the hierarchical recurrent network is proposed to encode each ERP signal and model them in each repetition with a temporal manner to predict which visual event elicited an ERP, while adversarial training is useful to capture more properties of stochasticity acting the same as increasing training data.
We conduct our experiments on one published visual ERP-based BCI task with 15 subjects and 3 different auditory workload states. The results demonstrate that our method can work effectively under different situations and achieve satisfactory performance compared with the only subject-specific baseline. In the specific-subject task and mixed-subject task, the averaged result of all the repetitions and states overpass the baseline by 8.50% and 8.96%, respectively, and even 1.39% is increased in the cross-subject task. Direct comparison between the performance of our model and the original baseline in specific-subject, mixed-subject and cross-subject tasks is as follows.
- The more direct comparison between the mixed-subject task and specific-subject task under all repetitions is reported here, as the supplementary material of our paper. It is easy to see that the results obtained in the mixed-subject task are even better than the results of the specific-subject task in many cases.
- There are two subjects performing not well in almost all tasks, here we report the performance of our model after removing these two "bad" subjects. The updated comparison of the results of 13 subjects with original grand-averaged results of 15 subjects are shown below.
- For demonstration, ERP waveforms in the six most relevant channels (‘PO3’, ‘POZ’, ‘PO4’, ‘O1’, ‘OZ’, ‘O2’) of some subjects (10) including two 'bad' subjects are also reported.
If you have the interest in our work, or use this code or part of it, please cite us!
Consider citing:
@article{ni2022bci,
author={Ni, Ziyi and Xu, Jiaming and Wu, Yuwei and Li, Mengfan and Xu, Guizhi and Xu, Bo},
title={Improving Cross-State and Cross-Subject Visual ERP-Based BCI With Temporal Modeling and Adversarial Training},
journal={IEEE Transactions on Neural Systems and Rehabilitation Engineering},
year={2022},
volume={30},
number={},
pages={369-379},
doi={10.1109/TNSRE.2022.3150007}}
For more detailed descirption, you can further explore the whole paper with this link.
Any other questions, feel free to contact us at niziyi2021@ia.ac.cn or jiaming.xu@ia.ac.cn