Skip to content

Pytorch codes of our paper "Improving Cross-State and Cross-Subject Visual ERP-based BCI with Temporal Modeling and Adversarial Training"

License

Notifications You must be signed in to change notification settings

aispeech-lab/VisBCI

Repository files navigation

VisBCI

Pytorch codes of our paper "Improving Cross-State and Cross-Subject Visual ERP-based BCI with Temporal Modeling and Adversarial Training", which is published in TNSRE.

A Simple Introduciton

Brain-computer interface (BCI) is a useful device for people without relying on peripheral nerves and muscles. However, the performance of the event-related potential (ERP)-based BCI declines when applying it to the real environments, especially in cross-state and cross-subject conditions.

For addressing this problem, we propose a general method with hierarchical temporal modeling and adversarial training to enhance the performance of the visual ERP-based BCI under different auditory workload states. The artitecture of our model is shown below.

The rationality of our method here is that the ERP-based BCI is based on electroencephalography (EEG) signals recorded from the surface of the scalp, continuously changing with time and somewhat stochastic. Therefore, the hierarchical recurrent network is proposed to encode each ERP signal and model them in each repetition with a temporal manner to predict which visual event elicited an ERP, while adversarial training is useful to capture more properties of stochasticity acting the same as increasing training data.

Result

We conduct our experiments on one published visual ERP-based BCI task with 15 subjects and 3 different auditory workload states. The results demonstrate that our method can work effectively under different situations and achieve satisfactory performance compared with the only subject-specific baseline. In the specific-subject task and mixed-subject task, the averaged result of all the repetitions and states overpass the baseline by 8.50% and 8.96%, respectively, and even 1.39% is increased in the cross-subject task. Direct comparison between the performance of our model and the original baseline in specific-subject, mixed-subject and cross-subject tasks is as follows.

Supplementary results

  1. The more direct comparison between the mixed-subject task and specific-subject task under all repetitions is reported here, as the supplementary material of our paper. It is easy to see that the results obtained in the mixed-subject task are even better than the results of the specific-subject task in many cases.
  1. There are two subjects performing not well in almost all tasks, here we report the performance of our model after removing these two "bad" subjects. The updated comparison of the results of 13 subjects with original grand-averaged results of 15 subjects are shown below.
  1. For demonstration, ERP waveforms in the six most relevant channels (‘PO3’, ‘POZ’, ‘PO4’, ‘O1’, ‘OZ’, ‘O2’) of some subjects (10) including two 'bad' subjects are also reported.
The ERP waveforms are for every 10 subjects of all the target stimuli in L state in 6 typical channels, with two ”bad” subjects (subject-6 and subject-10) highlighted. Two less obvious and standard curves are observed, which are difficult to recognize ERP waveforms in almost all the above channels. The ordinate represents the magnitude of the ERP amplitude and the abscissa represents the time after the target stimulation. Different subjects are represented by different colors.

Cite us

If you have the interest in our work, or use this code or part of it, please cite us!
Consider citing:

@article{ni2022bci,  
  author={Ni, Ziyi and Xu, Jiaming and Wu, Yuwei and Li, Mengfan and Xu, Guizhi and Xu, Bo},
  title={Improving Cross-State and Cross-Subject Visual ERP-Based BCI With Temporal Modeling and Adversarial Training},  
  journal={IEEE Transactions on Neural Systems and Rehabilitation Engineering},
  year={2022},
  volume={30},
  number={},
  pages={369-379},
  doi={10.1109/TNSRE.2022.3150007}}

For more detailed descirption, you can further explore the whole paper with this link.

Contact us

Any other questions, feel free to contact us at niziyi2021@ia.ac.cn or jiaming.xu@ia.ac.cn

About

Pytorch codes of our paper "Improving Cross-State and Cross-Subject Visual ERP-based BCI with Temporal Modeling and Adversarial Training"

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages