Assessing user bias in affect detection within context-based spoken dialog systems
Data(s) |
2012
|
---|---|
Resumo |
This paper presents an empirical evidence of user bias within a laboratory-oriented evaluation of a Spoken Dialog System. Specifically, we addressed user bias in their satisfaction judgements. We question the reliability of this data for modeling user emotion, focusing on contentment and frustration in a spoken dialog system. This bias is detected through machine learning experiments that were conducted on two datasets, users and annotators, which were then compared in order to assess the reliability of these datasets. The target used was the satisfaction rating and the predictors were conversational/dialog features. Our results indicated that standard classifiers were significantly more successful in discriminating frustration and contentment and the intensities of these emotions (reflected by user satisfaction ratings) from annotator data than from user data. Indirectly, the results showed that conversational features are reliable predictors of the two abovementioned emotions. |
Formato |
application/pdf |
Identificador | |
Idioma(s) |
eng |
Publicador |
E.T.S.I. Telecomunicación (UPM) |
Relação |
http://oa.upm.es/19784/1/INVE_MEM_2012_132557.pdf http://ieeexplore.ieee.org/xpls/abs_all.jsp?arnumber=6406341 |
Direitos |
http://creativecommons.org/licenses/by-nc-nd/3.0/es/ info:eu-repo/semantics/openAccess |
Fonte |
ASE/IEEE International Conference on Social Computing and 2012 ASE/IEEE International Conference on Privacy, Security, Risk and Trust | ASE/IEEE International Conference on Social Computing and 2012 ASE/IEEE International Conference on Privacy, Security, Risk and Trust | 03/09/2012 - 06/09/2012 | Amsterdam, The Netherlands |
Palavras-Chave | #Telecomunicaciones #Matemáticas #Psicología #Educación |
Tipo |
info:eu-repo/semantics/conferenceObject Ponencia en Congreso o Jornada PeerReviewed |