Please use this identifier to cite or link to this item:
Title: Engagement recognition using audio channel only
Author(s): Dresvyanskiy, Denis
Siegert, Ingo
Karpov, Alexei
Minker, Wolfgang
Issue Date: 2021
Type: Conference object
Language: English
URN: urn:nbn:de:gbv:ma9:1-1981185920-387212
Subjects: Paralinguistics
Human-computer interaction
Engagement recognition
Audio processing
Abstract: INTRODUCTION Utilizing dialogue assistants endowed with weak artificial intelligence has become a common technology, which is widespread across many industrial spheres - from operating robots using voice to speaking with an intelligent bot by telephone. However, such systems are still far from being essentially intelligent systems, since they cannot fully mimicry or replace humans during human-computer interaction (HCI). Nowadays, paralinguistic analyses is becoming one of the most important parts of HCI, because current requirements to such systems have been increased due to sharped improvement of speech-recognition systems: now, the HCI system should not only recognize, what the user is talking about, but also how he/she is talking, and which intention/state does he/she have now. Those include analyzing and evaluating such high-level features of dialogue as stress, emotions, engagement, and many others. Although there have been a lot of studies in paralinguistics devoted to recognizing high-level features (such as emotions[1] and stress[17, 25]) using audio cues, there are still almost no insights on how it could work for engagement.
Open Access: Open access publication
License: (CC BY-SA 4.0) Creative Commons Attribution ShareAlike 4.0(CC BY-SA 4.0) Creative Commons Attribution ShareAlike 4.0
Appears in Collections:Fakultät für Elektrotechnik und Informationstechnik (OA)

Files in This Item:
File Description SizeFormat 
AI-Debate2021_Dresvyanskiy et al._Final.pdfArtikel in Kongreßband484.63 kBAdobe PDFThumbnail