Martinez, H. P.2; Yannakakis, G. N.2; Hallam, John3
1 The Maersk Mc-Kinney Moller Institute, Faculty of Engineering, SDU2 unknown3 The Maersk Mc-Kinney Moller Institute, Faculty of Engineering, SDU
How should affect be appropriately annotated and how should machine learning best be employed to map manifestations of affect to affect annotations? What is the use of ratings of affect for the study of affective computing and how should we treat them? These are the key questions this paper attempts to address by investigating the impact of dissimilar representations of annotated affect on the efficacy of affect modelling. In particular, we compare several different binary-class and pairwise preference representations for automatically learning from ratings of affect. The representations are compared and tested on three datasets: one synthetic dataset (testing "in vitro") and two affective datasets (testing "in vivo"). The synthetic dataset couples a number of attributes with generated rating values. The two affective datasets contain physiological and contextual user attributes, and speech attributes, respectively; these attributes are coupled with ratings of various affective and cognitive states. The main results of the paper suggest that ratings (when used) should be naturally transformed to ordinal (ranked) representations for obtaining more reliable and generalisable models of affect. The findings of this paper have a direct impact on affect annotation and modelling research but, most importantly, challenge the traditional state-of-practice in affective computing and psychometrics at large.
I E E E Transactions on Affective Computing, 2014, Vol 5, Issue 3, p. 314-326
Affect annotation affect modelling ratings ranks preference learning classification computer games sensitive artificial listener (SAL) corpus MODELS RECOGNITION EMOTIONS