Title Emotion recognition in talking-face videos using persistent entropy and neural networks
Authors PALUZO HIDALGO, EDUARDO, Gonzalez-Diaz, Rocio , Aguirre-Carrazana, Guillermo
External publication Si
Means Electron. Res. Arch.
Scope Article
Nature Científica
JCR Quartile 3
SJR Quartile 3
JCR Impact 0.8
SJR Impact 0.314
Web https://www.scopus.com/inward/record.uri?eid=2-s2.0-85125926276&doi=10.3934%2fera.2022034&partnerID=40&md5=8a460f6df4aa9ff85be6110ee9c60826
Publication date 01/01/2022
ISI 000802224900014
Scopus Id 2-s2.0-85125926276
DOI 10.3934/era.2022034
Abstract The automatic recognition of a person\'s emotional state has become a very active research field that involves scientists specialized in different areas such as artificial intelligence, computer vision, or psychology, among others. Our main objective in this work is to develop a novel approach, using persistent entropy and neural networks as main tools, to recognise and classify emotions from talking-face videos. Specifically, we combine audio-signal and image-sequence information to compute a topology signature (a 9-dimensional vector) for each video. We prove that small changes in the video produce small changes in the signature, ensuring the stability of the method. These topological signatures are used to feed a neural network to distinguish between the following emotions: calm, happy, sad, angry, fearful, disgust, and surprised. The results reached are promising and competitive, beating the performances achieved in other state-of-the-art works found in the literature.
Keywords topological data analysis; persistent homology; persistent entropy; neural networks; audio-visual emotion recognition; talking-face videos
Universidad Loyola members