TY - JOUR
T1 - MixedEmotions
T2 - an open-source toolbox for multimodal emotion analysis
AU - Buitelaar, Paul
AU - Wood, Ian D.
AU - Negi, Sapna
AU - Arcan, Mihael
AU - McCrae, John P.
AU - Abele, Andrejs
AU - Robin, Cécile
AU - Andryushechkin, Vladimir
AU - Ziad, Housam
AU - Sagha, Hesam
AU - Schmitt, Maximilian
AU - Schuller, Björn W.
AU - Sánchez-Rada, J. Fernando
AU - Iglesias, Carlos A.
AU - Navarro, Carlos
AU - Giefer, Andreas
AU - Heise, Nicolaus
AU - Masucci, Vincenzo
AU - Danza, Francesco A.
AU - Caterino, Ciro
AU - Smrž, Pavel
AU - Hradis, Michal
AU - Povolný, Filip
AU - Klimeś, Marek
AU - Matějka, Pavel
AU - Tummarello, Giovanni
PY - 2018/9
Y1 - 2018/9
N2 - Recently, there is an increasing tendency to embed functionalities for recognizing emotions from user-generated media content in automated systems such as call-centre operations, recommendations, and assistive technologies, providing richer and more informative user and content profiles. However, to date, adding these functionalities was a tedious, costly, and time-consuming effort, requiring identification and integration of diverse tools with diverse interfaces as required by the use case at hand. The MixedEmotions Toolbox leverages the need for such functionalities by providing tools for text, audio, video, and linked data processing within an easily integrable plug-and-play platform. These functionalities include: 1) for text processing: emotion and sentiment recognition; 2) for audio processing: emotion, age, and gender recognition; 3) for video processing: face detection and tracking, emotion recognition, facial landmark localization, head pose estimation, face alignment, and body pose estimation; and 4) for linked data: knowledge graph integration. Moreover, the MixedEmotions Toolbox is open-source and free. In this paper, we present this toolbox in the context of the existing landscape, and provide a range of detailed benchmarks on standard test-beds showing its state-of-the-art performance. Furthermore, three real-world use cases show its effectiveness, namely, emotion-driven smart TV, call center monitoring, and brand reputation analysis.
AB - Recently, there is an increasing tendency to embed functionalities for recognizing emotions from user-generated media content in automated systems such as call-centre operations, recommendations, and assistive technologies, providing richer and more informative user and content profiles. However, to date, adding these functionalities was a tedious, costly, and time-consuming effort, requiring identification and integration of diverse tools with diverse interfaces as required by the use case at hand. The MixedEmotions Toolbox leverages the need for such functionalities by providing tools for text, audio, video, and linked data processing within an easily integrable plug-and-play platform. These functionalities include: 1) for text processing: emotion and sentiment recognition; 2) for audio processing: emotion, age, and gender recognition; 3) for video processing: face detection and tracking, emotion recognition, facial landmark localization, head pose estimation, face alignment, and body pose estimation; and 4) for linked data: knowledge graph integration. Moreover, the MixedEmotions Toolbox is open-source and free. In this paper, we present this toolbox in the context of the existing landscape, and provide a range of detailed benchmarks on standard test-beds showing its state-of-the-art performance. Furthermore, three real-world use cases show its effectiveness, namely, emotion-driven smart TV, call center monitoring, and brand reputation analysis.
KW - affective computing
KW - audio processing
KW - Emotion analysis
KW - linked data
KW - open source toolbox
KW - text processing
KW - video processing
UR - http://www.scopus.com/inward/record.url?scp=85040988311&partnerID=8YFLogxK
U2 - 10.1109/TMM.2018.2798287
DO - 10.1109/TMM.2018.2798287
M3 - Article
SN - 1520-9210
VL - 20
SP - 2454
EP - 2465
JO - IEEE Transactions on Multimedia
JF - IEEE Transactions on Multimedia
IS - 9
ER -