Show simple item record

Files in this item

Thumbnail

Item metadata

dc.contributor.authorMansouri Benssassi, Esma
dc.contributor.authorYe, Juan
dc.date.accessioned2021-02-11T10:30:11Z
dc.date.available2021-02-11T10:30:11Z
dc.date.issued2021-01-16
dc.identifier271472115
dc.identifiere9a21239-e0f8-4e9f-ae18-67f9b6b1063a
dc.identifier85100168720
dc.identifier000608096600002
dc.identifier.citationMansouri Benssassi , E & Ye , J 2021 , ' Generalisation and robustness investigation for facial and speech emotion recognition using bio-inspired spiking neural networks ' , Soft Computing , vol. First Online . https://doi.org/10.1007/s00500-020-05501-7en
dc.identifier.issn1432-7643
dc.identifier.otherORCID: /0000-0002-2838-6836/work/88731535
dc.identifier.urihttps://hdl.handle.net/10023/21409
dc.description.abstractEmotion recognition through facial expression and non verbal speech represent an important area in affective computing. They have been extensive studied, from classical feature extraction techniques to more recent deep learning approaches. However most of these approaches face two major challenges: (1) robustness – in the face of degradation such as noise, can a model still make correct predictions?, and (2) cross-dataset generalisation – when a model is trained on one dataset, can it be used to make inference on another dataset?. To directly address these challenges, we first propose the application of a Spiking Neural Network (SNN) in predicting emotional states based on facial expression and speech data, then investigate and compare their accuracy when facing data degradation or unseen new input. We evaluate our approach on third-party, publicly available datasets and compare to the state-of-the-art techniques. Our approach demonstrates robustness to noise, where it achieves an accuracy of 56.2% for facial expression recognition (FER) compared to 22.64% and 14.10% for CNN and SVM respectively when input images are degraded with the noise intensity of 0.5, and the highest accuracy of 74.3% for speech emotion recognition (SER) compared to 21.95% of CNN and 14.75% for SVM when audio white noise is applied. For generalisation, our approach achieves consistently high accu- racy of 89% for FER and 70% for SER in cross-dataset evaluation and suggests that it can learn more effective feature representations, which lead to good generalisa- tion of facial features and vocal characteristics across subjects.
dc.format.extent14
dc.format.extent2096514
dc.language.isoeng
dc.relation.ispartofSoft Computingen
dc.subjectSpiking neural networken
dc.subjectFacial emotion recognitionen
dc.subjectSpeech emotion recognitionen
dc.subjectUnsupervised learningen
dc.subjectQA75 Electronic computers. Computer scienceen
dc.subjectT Technologyen
dc.subjectNDASen
dc.subject.lccQA75en
dc.subject.lccTen
dc.titleGeneralisation and robustness investigation for facial and speech emotion recognition using bio-inspired spiking neural networksen
dc.typeJournal articleen
dc.contributor.institutionUniversity of St Andrews. School of Computer Scienceen
dc.identifier.doi10.1007/s00500-020-05501-7
dc.description.statusPeer revieweden
dc.date.embargoedUntil2021-01-16


This item appears in the following Collection(s)

Show simple item record