Improved BCI calibration in multimodal emotion recognition using heterogeneous adversarial transfer learning

View article
PeerJ Computer Science

Main article text

 

Introduction

Heterogeneous adversarial transfer learning framework for emotion recognition

System architecture

  • Generator: This module is responsible for generating synthetic EEG signals from non-EEG sensory data. It is trained using calibration data from various training subjects.

  • Discriminator: The discriminator evaluates the synthetic EEG signals to determine their authenticity. It provides feedback to improve the generator’s performance.

  • Trained generator: This module generates synthetic EEG signals using non-EEG sensory data from test subjects.

  • Classifiers for emotion recognition: These classifiers are trained to recognize emotions based on the generated EEG signals with non-EEG sensory data.

  • Trained classifiers for emotion recognition: These classifiers predicts the emotional state of test subjects using a combination of raw EEG data and non-EEG sensory data.

Operational phases

Training phase

Calibration phase

Testing phase

Generative adversarial network models

Emotion recognition models

Datasets and extracted features

SEED-V

DEAP

GraffitiVR

Experiments

Hardware and software

HATL parameters

Network architectures

  • SEED-V: Layers with 64, 128, and 256 units.

  • DEAP: Layers with 128 and 128 units.

  • GraffitiVR: Layers with 64 and 64 units.

  • SEED-V: Layers with 128 and 64 units.

  • DEAP: Layers with 64 and 64 units.

  • GraffitiVR: Layers with 32 and 32 units.

Implementation details

Computational complexity considerations for HATL

Evaluation criteria

Data similarity metrics

Classification performance metrics

Calibration improvement score

Experimental framework

Results

Performance of EEG data generation

Emotion recognition results using different GAN models

Emotion recognition results using SEED-V dataset

Emotion recognition results using DEAP dataset

Emotion recognition results using GraffitiVR dataset

Results about calibration duration using GraffitiVR dataset

Discussion

Conclusions

Limitations and future work

Supplemental Information

Code scripts for preprocessing and applying different GAN architectures to datasets.

Directory Structure codes/ Deap/ HATL/ : Contains code scripts related to the Heterogeneous Adversarial Transfer Learning (HATL) model for the DEAP dataset. preprocess/ : Contains preprocessing scripts for the DEAP dataset. GraffitiVR/ HATL/ : Contains code scripts related to the Heterogeneous Adversarial Transfer Learning (HATL) model for the GraffitiVR dataset. lucas_kanade/ : Contains scripts utilizing the Lucas-Kanade method for the GraffitiVR dataset. SeedV/ HATL/ : Contains code scripts related to the Heterogeneous Adversarial Transfer Learning (HATL) model for the SEED-V dataset.

DOI: 10.7717/peerj-cs.2649/supp-1

Raw performance data 1.

Performance metrics for various methods applied on the SEED-V, DEAP and GraffitiVR dataset. Each row represents a different method, including traditional modalities (e.g., Non-EEG Sensory Data), Generative Adversarial Networks (GANs) like CGAN and CWGAN, and combined EEG and non-EEG sensory data approaches. The metrics include average accuracy, standard deviation of accuracy, average precision, precision standard deviation, average recall, recall standard deviation, along with specific accuracies and standard deviations for Support Vector Machines (SVM), Random Forest (RF), and Multi-Layer Perceptron (MLP) classifiers. Each sheet shows related dataset results. Last sheet, titled “Calibration,” displays performance metrics related to the calibration duration required for different data modalities in emotion recognition tasks using the GraffitiVR dataset.

DOI: 10.7717/peerj-cs.2649/supp-2

Raw performance data 2.

Raw performance dataset provides performance metrics for various methods applied to the SEED-V, DEAP, and GraffitiVR datasets. Metrics include Euclidean and Wasserstein distances, as well as KL Divergence, Wasserstein Distance, T Statistic, and T p-value for each EEG frequency band.

DOI: 10.7717/peerj-cs.2649/supp-3

Additional Information and Declarations

Competing Interests

Author Contributions

Data Availability

Funding

The authors received no funding for this work.

MIT

Your institution may have Open Access funds available for qualifying authors. See if you qualify

Publish for free

Comment on Articles or Preprints and we'll waive your author fee
Learn more

Five new journals in Chemistry

Free to publish • Peer-reviewed • From PeerJ
Find out more