Multi-Modal Emotion Aware System Based on Fusion of Speech and Brain Information
Date
2019
Journal Title
Journal ISSN
Volume Title
Publisher
MDPI
Abstract
: In multi-modal emotion aware frameworks, it is essential to estimate the emotional features
then fuse them to different degrees. This basically follows either a feature-level or decision-level
strategy. In all likelihood, while features from several modalities may enhance the classification
performance, they might exhibit high dimensionality and make the learning process complex for the
most used machine learning algorithms. To overcome issues of feature extraction and multi-modal
fusion, hybrid fuzzy-evolutionary computation methodologies are employed to demonstrate
ultra-strong capability of learning features and dimensionality reduction. This paper proposes
a novel multi-modal emotion aware system by fusing speech with EEG modalities. Firstly, a mixing
feature set of speaker-dependent and independent characteristics is estimated from speech signal.
Further, EEG is utilized as inner channel complementing speech for more authoritative recognition,
by extracting multiple features belonging to time, frequency, and time–frequency. For classifying
unimodal data of either speech or EEG, a hybrid fuzzy c-means-genetic algorithm-neural network
model is proposed, where its fitness function finds the optimal fuzzy cluster number reducing the
classification error. To fuse speech with EEG information, a separate classifier is used for each
modality, then output is computed by integrating their posterior probabilities. Results show the
superiority of the proposed model, where the overall performance in terms of accuracy average
rates is 98.06%, and 97.28%, and 98.53% for EEG, speech, and multi-modal recognition, respectively.
The proposed model is also applied to two public databases for speech and EEG, namely: SAVEE and
MAHNOB, which achieve accuracies of 98.21% and 98.26%, respectively
Description
Keywords
multi-modal emotion aware systems; speech processing; EEG signal processing; hybrid
classification models
Citation
Rania M. Ghoniem, Abeer D. Algarni and Khaled Shaalan (2019) “Multi-Modal Emotion Aware System Based on Fusion of Speech and Brain Information,” Information, 10(7), p. 239.