Emotion Assessment for Affective Computing Based on Brain and ...

Emotion Assessment for Affective Computing Based on Brain and ... Emotion Assessment for Affective Computing Based on Brain and ...

theses.eurasip.org
from theses.eurasip.org More from this publisher
10.07.2015 Views

UNIVERSITE DE GENEVEDépartement d’Inong>forong>matiqueFACULTE DES SCIENCESProfesseur Thierry Punong>Emotionong> ong>Assessmentong> ong>forong> ong>Affectiveong> ong>Computingong>ong>Basedong> on Brain and Peripheral SignalsTHESEprésenté à la faculté des sciences de l’Université de Genèvepour obtenir le grade de Docteur ès sciences, mention inong>forong>matiqueparGuillaume ChaneldeCrozet (France)Thèse N° 4123HSE Print, Helsinki, Finland2009

UNIVERSITE DE GENEVEDépartement d’In<str<strong>on</strong>g>for</str<strong>on</strong>g>matiqueFACULTE DES SCIENCESProfesseur Thierry Pun<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> <str<strong>on</strong>g>Assessment</str<strong>on</strong>g> <str<strong>on</strong>g>for</str<strong>on</strong>g> <str<strong>on</strong>g>Affective</str<strong>on</strong>g> <str<strong>on</strong>g>Computing</str<strong>on</strong>g><str<strong>on</strong>g>Based</str<strong>on</strong>g> <strong>on</strong> <strong>Brain</strong> <strong>and</strong> Peripheral SignalsTHESEprésenté à la faculté des sciences de l’Université de Genèvepour obtenir le grade de Docteur ès sciences, menti<strong>on</strong> in<str<strong>on</strong>g>for</str<strong>on</strong>g>matiqueparGuillaume ChaneldeCrozet (France)Thèse N° 4123HSE Print, Helsinki, Finl<strong>and</strong>2009


La Faculté des sciences, sur le préavis de M<strong>on</strong>sieur Th. PUN, professeur ordinaire etdirecteur de thèse (Département d’in<str<strong>on</strong>g>for</str<strong>on</strong>g>matique), Madame M. PANTIC, professeure(Imperial College – L<strong>on</strong>d<strong>on</strong>, United Kingdom/University of Twente – The Netherl<strong>and</strong>s),Messieurs D. GRANDJEAN, professeur assistant (Secti<strong>on</strong> de psychologie), et S.VOLOSHYNOVSKIY, professeur adjoint suppléant (Département d’in<str<strong>on</strong>g>for</str<strong>on</strong>g>matique),autorise l’impressi<strong>on</strong> de la présente thèse, sans exprimer d’opini<strong>on</strong> sur les propositi<strong>on</strong>squi y s<strong>on</strong>t én<strong>on</strong>cées.Genève, le 28 juillet 2009Thèse - 4123 -


RemerciementsEffectuer une thèse de doctorat est un travail de l<strong>on</strong>gue haleine qui comporte de b<strong>on</strong>s côtéscomme l’apprentissage de nouveaux c<strong>on</strong>cepts, la découverte permanente ou les multiplesvoyages et c<strong>on</strong>férences, mais aussi des moments plus difficiles. Afin de surm<strong>on</strong>ter ces derniers, lesoutien des pers<strong>on</strong>nes qui vous entourent est indispensable. J’ai eu la chance d’avoir bénéficié decet appui et c’est avec gratitude que je m’adresse à toutes celles et ceux qui <strong>on</strong>t c<strong>on</strong>tribué àl’aboutissement de la présente recherche. Une thèse est aussi un travail de collaborati<strong>on</strong> et je suispersuadé que les pers<strong>on</strong>nes citées ci-dessous <strong>on</strong>t toutes c<strong>on</strong>tribué de manière importante aurésultat final.Tout d’abord, je tiens à remercier Thierry Pun, le directeur de cette thèse et du CVML, pour s<strong>on</strong>encadrement, l’expérience qu’il m’a transmise, les resp<strong>on</strong>sabilités qu’il m’a c<strong>on</strong>fiées et lac<strong>on</strong>fiance qu’il a bien voulu m’accorder. Thierry a su créer au sein de s<strong>on</strong> laboratoire uneambiance de travaille agréable qui favorise les échanges et stimule la recherche, ce qui agr<strong>and</strong>ement c<strong>on</strong>tribué à ce travail de thèse. Je témoigne également ma rec<strong>on</strong>naissance à StéphaneMach<strong>and</strong>-Maillet et Sviatoslav Voloshynovskiy pour m’avoir ouvert les portes de leurs groupes,pour leurs c<strong>on</strong>seils et de m’avoir poussé à l’excellence.Un des moments important d’une thèse est la soutenance finale, qui ne peut s’accomplir sans laprésence du jury et de leurs nombreuses questi<strong>on</strong>s. J’exprime tout ma gratitude aux professeursMaja Pantic, Didier Gr<strong>and</strong>jean, Sviatoslav Voloshynovskiy et Thierry Pun qui <strong>on</strong>t su jouer cerôle à merveille en rec<strong>on</strong>naissant les mérites de cette thèse et en discutant ses défauts de manièrec<strong>on</strong>structive. Je profite de ce paragraphe pour remercier tout particulièrement Didier Gr<strong>and</strong>jeanainsi que David S<strong>and</strong>er de m’avoir introduit au m<strong>on</strong>de de la psychologie et pour les multiplesdiscussi<strong>on</strong>s et suggesti<strong>on</strong>s qui n’<strong>on</strong>t pu qu’améliorer la qualité de ce travail.En outre, toute m<strong>on</strong> attenti<strong>on</strong> se porte vers les membres du CVML avec qui j’ai collaboré cesdernières années. Merci à Julien Kr<strong>on</strong>egg de m’avoir appris à relativiser, à Eric Bruno pour s<strong>on</strong>aide en matière de classificati<strong>on</strong>, à Benoît Deville pour s<strong>on</strong> soutien psychologique « BDisé », àMohammad Soleymani de m’avoir aidé à fouiller les d<strong>on</strong>nées ainsi que pour s<strong>on</strong> amitié, à JoepKierkels pour le thé intellectuel du matin et à tous pour votre aide, nos discussi<strong>on</strong>s endiablées, lesbarbecues, les repas, les sorties escalade et les soirées dansantes. J’ai également eu beaucoup deplaisir à travailler avec Mr. Et Mme. Hertzschuch ainsi qu’avec tous mes collègues de Soprotec,merci de m’avoir appris ce magnifique métier de <str<strong>on</strong>g>for</str<strong>on</strong>g>mateur.Le soutien moral et émoti<strong>on</strong>nel étant des plus importants dans les moments exigeants, je remerciechaleureusement Annina qui a eu la patience de supporter mes états d’âme pendant plusieursannées et qui a su me d<strong>on</strong>ner le courage nécessaire. Impossible d’oublier les « collocs »,iii


Séverine, Lydie, Nicolas et Sylvain. Que ce soit par la fête, grâce aux légumes et fleurs du jardinou par nos discussi<strong>on</strong>s, j’ai passé de très b<strong>on</strong>s moments en leurs compagnie qui <strong>on</strong>t enrichit cesannées de thèse. Mes pensées v<strong>on</strong>t également vers tous mes amis d’hier et d’aujourd’hui quic<strong>on</strong>stituent des sources d’inspirati<strong>on</strong>. Finalement, un gr<strong>and</strong> merci à mes parents pour leursencouragements et leur présence durant les 30 dernières années.iv


RésuméLes Interfaces Homme-Machine actuelles manquent « d’intelligence émoti<strong>on</strong>nelle » : elles nes<strong>on</strong>t pas capables d’identifier les émoti<strong>on</strong>s humaines et de prendre cette in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> en comptepour choisir les acti<strong>on</strong>s à exécuter. Le but de l’in<str<strong>on</strong>g>for</str<strong>on</strong>g>matique affective ou affective computing estde combler ce manque en détectant les indices émoti<strong>on</strong>nels se produisant durant l’interacti<strong>on</strong>avec la machine et en synthétisant les rép<strong>on</strong>ses émoti<strong>on</strong>nelles adéquates. Durant ces dernièresannées, la plupart des études s’intéressant à la détecti<strong>on</strong> d’émoti<strong>on</strong>s se s<strong>on</strong>t c<strong>on</strong>centrées surl’analyse des expressi<strong>on</strong>s faciales et de la parole pour déterminer l’état émoti<strong>on</strong>nel d’unepers<strong>on</strong>ne. Toutefois, il existe d’autres sources d’in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> émoti<strong>on</strong>nelle provenant enparticulier de l'analyse des signaux physiologiques. Comparés à l’analyse des expressi<strong>on</strong>sfaciales, ces signaux n’<strong>on</strong>t été que peu étudiés bien qu’ils aient plusieurs avantages; par exemple,il est plus difficile de simuler une rép<strong>on</strong>se physiologique qu’une expressi<strong>on</strong> faciale.Cette thèse traite de l’utilisati<strong>on</strong> de deux types d’activités physiologiques pour détecter lesémoti<strong>on</strong>s dans le cadre de l’affective computing : l’activité du système nerveux central (lecerveau) et l’activité du système nerveux périphérique. L’activité centrale est mesurée parl’utilisati<strong>on</strong> d’électro-encéphalogrammes (EEGs). L’activité périphérique est mesurée au moyendes capteurs suivants : un capteur de rép<strong>on</strong>se cutané galvanique (GSR) permettant de mesurer latranspirati<strong>on</strong> ; une ceinture de respirati<strong>on</strong> mesurant les mouvements de l’abdomen ; unpléthysmographe enregistrant les variati<strong>on</strong>s de volume sanguin (blood volume pulse -BVP), et uncapteur de température servant à mesurer la température cutanée du doigt.L’espace valence-arousal (hed<strong>on</strong>icité-excitati<strong>on</strong>) est choisi pour représenter les émoti<strong>on</strong>s car,issu de la théorie cognitive des émoti<strong>on</strong>s, il est suffisamment général pour être utilisable pourplusieurs applicati<strong>on</strong>s. Certaines régi<strong>on</strong>s de l’espace valence-arousal s<strong>on</strong>t utilisées pour définirdes classes émoti<strong>on</strong>nelles. Afin de rec<strong>on</strong>naître ces classes à partir des signaux physiologiques, ilest nécessaire de trouver un modèle in<str<strong>on</strong>g>for</str<strong>on</strong>g>matique qui associe une classe à une activitéphysiologique d<strong>on</strong>née. Les algorithmes de rec<strong>on</strong>naissance de <str<strong>on</strong>g>for</str<strong>on</strong>g>me et d’apprentissage artificiels<strong>on</strong>t utilisés pour déterminer ce modèle.Trois protocoles s<strong>on</strong>t c<strong>on</strong>çus pour enregistrer les signaux physiologiques lors de stimulati<strong>on</strong>sémoti<strong>on</strong>nelles de différents types: regarder des images, se remémorer des épisodes émoti<strong>on</strong>nelspassés, et jouer à un jeu vidéo à différents niveaux de difficulté. Pour chaque stimulati<strong>on</strong>, descaractéristiques s<strong>on</strong>t extraites des signaux EEG et périphériques. Les signaux EEG s<strong>on</strong>tcaractérisés par leur énergie dans des b<strong>and</strong>es de fréquences en relati<strong>on</strong> avec les processusémoti<strong>on</strong>nels. De plus, l’in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> mutuelle calculée entre chaque paire d’électrodes estproposée comme un nouvel ensemble de caractéristiques pour détecter les émoti<strong>on</strong>s. Pour ce quiest des signaux périphériques, les caractéristiques s<strong>on</strong>t calculées sur la base de la littératurev


existante et discutées par rapport aux aspects temporels. Plusieurs classifieurs (naïve Bayes,analyse discriminante, machines à vecteurs de support, relevance vector machines) s<strong>on</strong>t ensuiteentraînés sur la base de d<strong>on</strong>nées des caractéristiques. La per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance des ces algorithmes estévaluée par rapport à leur précisi<strong>on</strong> pour des classificati<strong>on</strong>s intra- (deux premiers protocoles) etinter- (troisième protocole) participants. Des méthodes de sélecti<strong>on</strong> de caractéristiques s<strong>on</strong>temployées pour trouver les caractéristiques les plus intéressantes en vue de la détecti<strong>on</strong>d’émoti<strong>on</strong>s ainsi que pour réduire la taille de l’espace des caractéristiques. Finalement, la fusi<strong>on</strong>des in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>s du système nerveux central et périphérique est analysée tant au niveau descaractéristiques (avant classificati<strong>on</strong>) qu’au niveau décisi<strong>on</strong>nel (après classificati<strong>on</strong>).Les résultats m<strong>on</strong>trent que les signaux EEG s<strong>on</strong>t utilisables pour détecter les émoti<strong>on</strong>s car laprécisi<strong>on</strong> obtenue par les algorithmes de classificati<strong>on</strong> sur les d<strong>on</strong>nées EEG est bien supérieure àla précisi<strong>on</strong> aléatoire. La meilleure précisi<strong>on</strong> moyenne obtenue pour rec<strong>on</strong>naître trois classes estde 68% et autour de 80% pour deux classes (moyennes de la précisi<strong>on</strong> pour plusieursparticipants). Les caractéristiques calculées en utilisant l’in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> mutuelle entre les pairesd’électrodes s<strong>on</strong>t également utilisables pour détecter les émoti<strong>on</strong>s comme le m<strong>on</strong>trent lesrésultats obtenus (par exemple 62% de précisi<strong>on</strong> pour rec<strong>on</strong>naître trois états émoti<strong>on</strong>nels). Deplus la précisi<strong>on</strong> de classificati<strong>on</strong> en utilisant les signaux EEG est supérieure à celle obtenue avecles signaux périphériques lorsque leurs caractéristiques s<strong>on</strong>t calculées sur des fenêtrestemporelles relativement courtes (6 à 30 sec<strong>on</strong>des). La fusi<strong>on</strong> des in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>s périphériques etcentrales au niveau décisi<strong>on</strong>nel augmente significativement la précisi<strong>on</strong> de détecti<strong>on</strong> (de 3% à7%), ce qui encourage la fusi<strong>on</strong> avec d’autres sources d’in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> émoti<strong>on</strong>nelle comme lesexpressi<strong>on</strong>s faciales ou la parole. Une applicati<strong>on</strong> des méthodes développées est proposée commeexemple d’une interface affective. Celle-ci adapterait automatiquement le niveau de difficultéd’un jeu vidéo en f<strong>on</strong>cti<strong>on</strong> de l’état émoti<strong>on</strong>nel du joueur.vi


SummaryCurrent Human-Machine Interfaces (HMI) lack of “emoti<strong>on</strong>al intelligence”, i.e. they are not ableto identify human emoti<strong>on</strong>al states <strong>and</strong> take this in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> into account to decide <strong>on</strong> the properacti<strong>on</strong>s to execute. The goal of affective computing is to fill this lack by detecting emoti<strong>on</strong>al cuesoccurring during Human-Computer Interacti<strong>on</strong> (HCI) <strong>and</strong> synthesizing emoti<strong>on</strong>al resp<strong>on</strong>ses. Inthe last decades, most of the studies <strong>on</strong> emoti<strong>on</strong> assessment have focused <strong>on</strong> the analysis of facialexpressi<strong>on</strong>s <strong>and</strong> speech to determine the emoti<strong>on</strong>al state of a pers<strong>on</strong>. Physiological activity alsoincludes emoti<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> that can be used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment but has received lessattenti<strong>on</strong> despite of its advantages (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance it can be less easily faked than facialexpressi<strong>on</strong>s).This thesis reports <strong>on</strong> the use of two types of physiological activities to assess emoti<strong>on</strong>s in thec<strong>on</strong>text of affective computing: the activity of the central nervous system (i.e. the brain) <strong>and</strong> theactivity of the peripheral nervous system. The central activity is m<strong>on</strong>itored by recording electroencephalograms(EEGs). The peripheral activity is assessed by using the following sensors: aGalvanic Skin Resp<strong>on</strong>se (GSR) sensor to measure sudati<strong>on</strong>; a respirati<strong>on</strong> belt to measureabdomen expansi<strong>on</strong>; a plethysmograph to record blood volume pulse (BVP); <strong>and</strong> a temperaturesensor to record finger temperature.The valence-arousal space is chosen to represent emoti<strong>on</strong>s since it originates from the cognitivetheory of emoti<strong>on</strong>s <strong>and</strong> is general enough to be usable <str<strong>on</strong>g>for</str<strong>on</strong>g> several applicati<strong>on</strong>s. Several areas inthe valence-arousal space are used as ground-truth classes. In order to automatically detect thoseclasses from physiological signals, it is necessary to find a computati<strong>on</strong>al model that maps agiven physiological pattern to <strong>on</strong>e of the chosen classes. Pattern recogniti<strong>on</strong> <strong>and</strong> machinelearning algorithms are employed to infer such a model.Three protocols that use different emoti<strong>on</strong> elicitati<strong>on</strong> methods (images, recall of past emoti<strong>on</strong>alepisodes <strong>and</strong> playing a video game at different difficulty levels) are designed to gatherphysiological signals during emoti<strong>on</strong>al stimulati<strong>on</strong>s. For each emoti<strong>on</strong>al stimulati<strong>on</strong>s, features areextracted from the EEG <strong>and</strong> peripheral signals. For EEG signals, energy features that are knownto be related to emoti<strong>on</strong>al processes are computed. Moreover, the Mutual In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> (MI)computed between all pairs of electrodes is proposed as a new set of features. For peripheralsignals, the computed features are chosen based <strong>on</strong> the review of the literature <strong>and</strong> are discussedrelatively to temporal aspects. Several classifiers (Naïve Bayes, Discriminant Analysis, SupportVector Machines – SVM <strong>and</strong> Relevance Vector Machines - RVM) are then trained <strong>on</strong> theresulting databases. The per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of these algorithms is evaluated according to the obtainedaccuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> both intra (two first protocols) <strong>and</strong> inter (third protocol) participant classificati<strong>on</strong>.Feature selecti<strong>on</strong> methods are also employed to find the most relevant features <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>vii


assessment <strong>and</strong> reduce the size of the original feature spaces. Finally, fusi<strong>on</strong> of the central <strong>and</strong>peripheral in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> at the decisi<strong>on</strong> <strong>and</strong> feature level is analyzed.The results show that EEG signals are usable <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment since the classificati<strong>on</strong>accuracy obtained with EEG features is much higher than the r<strong>and</strong>om level. The best accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g>the detecti<strong>on</strong> of three emoti<strong>on</strong>al classes is 68% <strong>and</strong> around 80% <str<strong>on</strong>g>for</str<strong>on</strong>g> two classes, when averagedacross participants. The effectiveness of the MI feature set <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment is alsodem<strong>on</strong>strated by the classificati<strong>on</strong> accuracies (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance 62% to detect three emoti<strong>on</strong>al classes).Moreover, the accuracy obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> based <strong>on</strong> EEG features is found to be higherthan based <strong>on</strong> peripheral features when the features are computed <strong>on</strong> short time periods (6 to 30sec<strong>on</strong>ds). Fusi<strong>on</strong> of the peripheral <strong>and</strong> EEG features at the decisi<strong>on</strong> level significantly increasedthe accuracy (by an amount of 3% to 7%), encouraging further fusi<strong>on</strong> with other sources ofemoti<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> (facial expressi<strong>on</strong>s, speech, etc.). An applicati<strong>on</strong> of the developed methodswhich automatically adapts the difficulty of games based <strong>on</strong> emoti<strong>on</strong> assessment is proposed asan example of affective HMI.viii


Table of c<strong>on</strong>tentsRésuméSummaryTable of c<strong>on</strong>tentsTable of symbolsAcr<strong>on</strong>ymsiiiviiixxiiixivChapter 1 Introducti<strong>on</strong> 11.1 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s <strong>and</strong> machines 11.2 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment 21.2.1 Multimodal expressi<strong>on</strong> of emoti<strong>on</strong> 21.2.2 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment as a comp<strong>on</strong>ent of HCI 41.2.3 Applicati<strong>on</strong>s of emoti<strong>on</strong> assessment 51.2.4 Related questi<strong>on</strong>s / problems 91.3 C<strong>on</strong>tributi<strong>on</strong>s 111.4 Thesis overview 11Chapter 2 State of the art 132.1 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> representati<strong>on</strong>s <strong>and</strong> models 132.1.1 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> categories <strong>and</strong> basic emoti<strong>on</strong>s 142.1.2 C<strong>on</strong>tinuous representati<strong>on</strong>s 162.1.3 Models of emoti<strong>on</strong>s 192.1.4 Finding an adequate representati<strong>on</strong> 242.2 Physiological signals 262.2.1 Central nervous system (CNS) 272.2.2 Peripheral nervous system (PNS) 332.2.3 Variability of physiological resp<strong>on</strong>ses 402.3 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment from physiological signals 41Chapter 3 Physiological signals recording <strong>and</strong> processing 513.1 Material 513.1.1 EEG 523.1.2 Peripheral sensors 53ix


3.2 Signals acquisiti<strong>on</strong> <strong>and</strong> preprocessing 553.2.1 Signal acquisiti<strong>on</strong> 553.2.2 Denoising 553.2.3 EEG re-referencing 563.2.4 HR computati<strong>on</strong> 563.3 Characterizati<strong>on</strong> of physiological activity 593.3.1 St<strong>and</strong>ard features 593.3.2 Advanced features 603.4 Ethical <strong>and</strong> privacy aspects 68Chapter 4 Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment 714.1 Classificati<strong>on</strong> 714.1.1 Ground-truth definiti<strong>on</strong> 714.1.2 Validati<strong>on</strong> strategies 744.1.3 Classifiers 764.2 Feature selecti<strong>on</strong> 804.2.1 Filter methods 814.2.2 The SFFS wrapper method 834.3 Fusi<strong>on</strong> 844.3.1 Feature level 854.3.2 Classifier level 854.4 Rejecti<strong>on</strong> of samples 86Chapter 5 <str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s elicited by visual stimuli 895.1 Introducti<strong>on</strong> 895.2 Data collecti<strong>on</strong> 895.2.1 Visual stimuli 895.2.2 Acquisiti<strong>on</strong> protocol 915.2.1 Features extracted 925.3 Classificati<strong>on</strong> 935.3.1 Ground-truth definiti<strong>on</strong>s 935.3.2 Methods 965.4 Results 975.4.1 Valence experiment 975.4.2 Arousal experiment 985.5 C<strong>on</strong>clusi<strong>on</strong> 100x


Chapter 6 <str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>s 1016.1 Introducti<strong>on</strong> 1016.2 Data acquisiti<strong>on</strong> 1016.2.1 Acquisiti<strong>on</strong> protocol 1016.2.2 Feature extracti<strong>on</strong> 1046.3 Classificati<strong>on</strong> 1056.3.1 The different classificati<strong>on</strong> schemes 1056.3.2 Classifiers 1066.3.3 Reducti<strong>on</strong> of the feature space 1076.3.4 Fusi<strong>on</strong> <strong>and</strong> rejecti<strong>on</strong> 1086.4 Results 1086.4.1 Participants reports <strong>and</strong> protocol validati<strong>on</strong> 1086.4.2 Results of single classifiers 1096.4.3 Results of feature selecti<strong>on</strong> 1146.4.4 Results of fusi<strong>on</strong> 1166.4.5 Results of rejecti<strong>on</strong> 1186.5 C<strong>on</strong>clusi<strong>on</strong>s 119Chapter 7 <str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer games 1217.1 Introducti<strong>on</strong>: the flow theory <str<strong>on</strong>g>for</str<strong>on</strong>g> games 1217.2 Data acquisiti<strong>on</strong> 1227.2.1 Acquisiti<strong>on</strong> protocol 1227.2.2 Feature extracti<strong>on</strong> 1257.3 Analysis of questi<strong>on</strong>naires <strong>and</strong> of physiological features 1277.3.1 Elicited emoti<strong>on</strong>s 1277.3.2 Evoluti<strong>on</strong> of emoti<strong>on</strong>s in engaged trials 1307.4 Classificati<strong>on</strong> of the gaming c<strong>on</strong>diti<strong>on</strong>s using physiological signals 1317.4.1 Classificati<strong>on</strong> methods 1317.4.2 Peripheral signals 1327.4.3 EEG signals 1357.4.4 EEG <strong>and</strong> peripheral signals 1377.4.5 Fusi<strong>on</strong> 1387.5 Analysis of game-over events 1397.5.1 Method 1407.5.2 Results 1407.6 C<strong>on</strong>clusi<strong>on</strong> 142xi


Chapter 8 C<strong>on</strong>clusi<strong>on</strong>s 1458.1 Outcomes 1458.2 Future prospects 147Appendix A C<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m 151Appendix B Neighborhood table <str<strong>on</strong>g>for</str<strong>on</strong>g> the Laplacian filter 155Appendix C List of IAPS images used 157Appendix D Questi<strong>on</strong>naire results <str<strong>on</strong>g>for</str<strong>on</strong>g> the game protocol 159Appendix E Publicati<strong>on</strong>s 161References 165List of figures 175List of tables 179xii


Table of symbolsx(n)xxxValue of the signal x at discrete time nMean of signal xSt<strong>and</strong>ard deviati<strong>on</strong> of signal xMean of the derivative of signal xMin , Max Minimum <strong>and</strong> maximum values of signal xxxNN eN sT2f 1Ff if iFCiNumber of trials <strong>and</strong> number of instances in the feature set FNumber of EEG electrodesNumber of samples in <strong>on</strong>e of the signals of a trialDurati<strong>on</strong> of a trialFeature having name <strong>and</strong> computed from signal Size of the feature space, number of features in a feature vector f i2Feature vector c<strong>on</strong>taining the values of some f 1features <str<strong>on</strong>g>for</str<strong>on</strong>g> the trial i(line of the feature set F)Vector c<strong>on</strong>taining the values of the feature i <str<strong>on</strong>g>for</str<strong>on</strong>g> several trials (column of thefeature set F)Matrix c<strong>on</strong>taining all the f i instancesNumber of classesClass iy i True label associated to trial i, y { 1, , }y ˆiEstimated label of trial i, yˆ i{ 1, , C}iCyŷColumn vector of true labels y iColumn vector of estimated labels y ˆixiii


Acr<strong>on</strong>ymsANOVAANSBCIBDIBPBVPCNSDBPECoGEDAEDREEGEMGFCBFFFTfMRIfNIRSGSRHCIHFHMIHRHRVIADSIAPSKNNLDALFMEGMIMSEANalysis Of VArianceAut<strong>on</strong>omic Nervous System<strong>Brain</strong>-Computer InterfaceBelief, Desire <strong>and</strong> Intensi<strong>on</strong>sBlood PressureBlood Volume PulseCentral Nervous SystemDiastolic Blood PressureElectro-Cortico-GraphyElectrodermal ActivityElectrodermal Resp<strong>on</strong>seElectro-encephalogramElectromyogramFast Correlati<strong>on</strong> <str<strong>on</strong>g>Based</str<strong>on</strong>g> FilterFast Fourier Trans<str<strong>on</strong>g>for</str<strong>on</strong>g>mfuncti<strong>on</strong>al Magnetic Res<strong>on</strong>ance Imageryfuncti<strong>on</strong>al Near InfraRed SpectroscopyGalvanic Skin Resp<strong>on</strong>seHuman-Computer Interacti<strong>on</strong>High Frequency (frequency b<strong>and</strong> of interest in the HR power spectrum)Human-Machine InterfacesHeart RateHeart Rate VariabilityInternati<strong>on</strong>al <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Digitized Sound systemInternati<strong>on</strong>al <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Picture SystemK-Nearest-NeighboorLinear Discriminant AnalysisLow Frequency (frequency b<strong>and</strong> of interest in the HR power spectrum)Magneto-encephalogramMutual In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>Mean Square Errorxiv


OCCPETPNSQDARVMSAMSBPSECSFFSSNSSPECTSTFTSVMVLFOrt<strong>on</strong>y, Clore <strong>and</strong> CollinsPositr<strong>on</strong> Emissi<strong>on</strong> TomographyPeripheral Nervous SystemQuadratic Discriminant AnalysisRelevance Vector MachineSelf <str<strong>on</strong>g>Assessment</str<strong>on</strong>g> ManikinSystolic Blood PressureSequential Evaluati<strong>on</strong> CheckSequential Floating Forward Selecti<strong>on</strong>Somatic Nervous SystemSingle Phot<strong>on</strong> Emissi<strong>on</strong> Computed TomographyShort-Time Fourier Trans<str<strong>on</strong>g>for</str<strong>on</strong>g>mSupport Vector MachineVery Low Frequency (frequency b<strong>and</strong> of interest in the HR power spectrum)xv


Introducti<strong>on</strong>Chapter 1Introducti<strong>on</strong>1.1 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s <strong>and</strong> machines<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s are part of any natural communicati<strong>on</strong> involving humans. They can be expressed eitherverbally through emoti<strong>on</strong>al vocabulary, or by expressing n<strong>on</strong>-verbal cues such as int<strong>on</strong>ati<strong>on</strong> ofvoice, facial expressi<strong>on</strong>s <strong>and</strong> gestures. In this c<strong>on</strong>text, decoding emoti<strong>on</strong>al cues is essential tocorrectly interpret a message. For instance, aggressively shouting “Shut up” at some<strong>on</strong>e does nothave the same meaning as saying it while laughing <strong>and</strong> smiling. In the sec<strong>on</strong>d case, <strong>on</strong>e couldinterpret that the speaker does not really want his interlocutor to shut up but rather that he foundhis remark funny <strong>and</strong> at the same time at the limit of social acceptance. However, disambiguating<strong>and</strong> enriching communicati<strong>on</strong> is not the <strong>on</strong>ly role of emoti<strong>on</strong>s. It has also been found that theyplay a key role in decisi<strong>on</strong> making <strong>and</strong> that they are str<strong>on</strong>gly related to tendencies to acti<strong>on</strong>. Thisdiscovery is certainly a reas<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the importance they have nowadays in neuromarketing thataim at uncovering the cerebral mechanisms leading to the purchase decisi<strong>on</strong>. In his bestsellerbook, D. Goleman [1] does not hesitate to talk about “emoti<strong>on</strong>al intelligence”, i.e. the ability ofsome<strong>on</strong>e to underst<strong>and</strong>, communicate <strong>and</strong> manage his / her own emoti<strong>on</strong>s as well as his / hercapacity to underst<strong>and</strong> <strong>and</strong> feel emoti<strong>on</strong>s of the others. According to him those competences areas important as logic <strong>and</strong> verbal skills to “succeed in life”. Of course the definiti<strong>on</strong> of what is“success in life” could be discussed, but nevertheless, the success of Goleman’s book <strong>and</strong> ofmany other books <strong>on</strong> emoti<strong>on</strong>s clearly dem<strong>on</strong>strates the importance that people attach to thistopic in our society.Despite the importance of emoti<strong>on</strong>s in communicati<strong>on</strong>, many Human-Machine Interfaces (HMI)completely lack “emoti<strong>on</strong>al intelligence”. The reacti<strong>on</strong> of users in resp<strong>on</strong>se to this shortage iseither to be more <strong>and</strong> more frustrated by the machine or to invent new ways to communicate theiremoti<strong>on</strong>s using the limited interfaces at their dispositi<strong>on</strong>. The number of internet videos showinga pers<strong>on</strong> getting upset in fr<strong>on</strong>t of his / her computer is a clear indicator of the first reacti<strong>on</strong>, whilethe use of emotic<strong>on</strong>s in chatting sessi<strong>on</strong>s <strong>and</strong> mails is an example of the sec<strong>on</strong>d. In the lastdecades, researchers <strong>and</strong> manufacturers started to take into account the emoti<strong>on</strong>s of users in thedesign of products, adding the c<strong>on</strong>cept of user experience to the <strong>on</strong>e of usability. For computersthis is part of what is called “human-centered computing”, where the goal is to have interfacesthat are made <str<strong>on</strong>g>for</str<strong>on</strong>g> the human: they should be intuitive, easily h<strong>and</strong>led, <strong>and</strong> use as most as possiblethe st<strong>and</strong>ard communicati<strong>on</strong> channels used by humans (speech, gestures, etc.). Current interfacesare far from this objective since they are complex <strong>and</strong> require training to be used. Even ifincluding the user experience in the design of software is a step toward the improvement of HMI,it is still far from having machines that can react properly to a sudden emoti<strong>on</strong>al reacti<strong>on</strong> of theuser.1


Chapter 1In order to obtain such an “emoti<strong>on</strong>al machine” it is necessary to include emoti<strong>on</strong>s in the humanmachinecommunicati<strong>on</strong> loop (see Figure 1.1). This is what is defined as “affective computing”.According to Picard [2], affective computing “proposes to give computers the ability to recognize[<strong>and</strong>] express […] emoti<strong>on</strong>s”. Synthetic expressi<strong>on</strong> of emoti<strong>on</strong>s can be achieved by enablingavatars or simpler agents to have facial expressi<strong>on</strong>s, different t<strong>on</strong>es of voice, <strong>and</strong> empathicbehaviors [3, 4]. Detecti<strong>on</strong> of human emoti<strong>on</strong>s can be realized by m<strong>on</strong>itoring <strong>and</strong> interpreting thedifferent cues that are given in both verbal <strong>and</strong> n<strong>on</strong>-verbal communicati<strong>on</strong>. However, a systemthat is “emoti<strong>on</strong>ally intelligent” should not <strong>on</strong>ly detect <strong>and</strong> express emoti<strong>on</strong>s but it should alsotake the proper acti<strong>on</strong> in resp<strong>on</strong>se to the detected emoti<strong>on</strong>. Expressing the adequate emoti<strong>on</strong> isthus <strong>on</strong>e of the outputs of this decisi<strong>on</strong> making. The proper acti<strong>on</strong> to take is dependent <strong>on</strong> theapplicati<strong>on</strong> but examples of reacti<strong>on</strong>s could be to provide help in the case the user feels helplessor to lower task dem<strong>and</strong> in the case he / she is highly stressed. Many applicati<strong>on</strong>s are detailed inchapter 1.2.2. Accurately assessing emoti<strong>on</strong>s is thus a critical step toward affective computingsince this will determine the reacti<strong>on</strong> of the system. The present work will focus <strong>on</strong> this first stepof affective computing by trying to reliably assess emoti<strong>on</strong> from several emoti<strong>on</strong>al cues.<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al cuesSensesSensorsOuput devices<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>assessmentDecisi<strong>on</strong> ofproper reacti<strong>on</strong><str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>synthesisUserMachine1.2 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessmentFigure 1.1. Including emoti<strong>on</strong>s in the human-machine loop1.2.1 Multimodal expressi<strong>on</strong> of emoti<strong>on</strong>A modality is defined as a path used to carry in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of interacti<strong>on</strong>. In HMI,there are two possible approaches to define a modality: from the machine point of view <strong>and</strong> fromthe user point of view. On the machine side, a modality refers to the processes that generatein<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> to the physical world <strong>and</strong> interpret in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> from it. It is thus possible todistinguish between input <strong>and</strong> output modalities <strong>and</strong> to associate them to the corresp<strong>on</strong>dingcommunicati<strong>on</strong> device. A keyboard, a mouse <strong>and</strong> a pad with the associated in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> processesare typical input modalities, while a text, an image <strong>and</strong> music presented <strong>on</strong> screens <strong>and</strong> speakers2


Introducti<strong>on</strong>are typical output modalities. From the user perspective, a modality corresp<strong>on</strong>ds to <strong>on</strong>e of thesenses used <str<strong>on</strong>g>for</str<strong>on</strong>g> interpreting (resp. transmitting) messages from (resp. to) the machine. Inpsychology a modality is similarly defined as <strong>on</strong>e of the five sensorial categories: touch, visi<strong>on</strong>,hearing, taste <strong>and</strong> smell. In this thesis, we choose to use the word modality in a wide sense,including both of those aspects.<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s can be expressed trough several channels <strong>and</strong> modalities. Facial expressi<strong>on</strong>s, gestures,postures, speech <strong>and</strong> int<strong>on</strong>ati<strong>on</strong> of voice have already been cited in the introducti<strong>on</strong> <strong>and</strong> arecertainly those that are the most obvious. However, emoti<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> can also be found inmany other modalities. For instance it has been shown that there are different physiologicalactivati<strong>on</strong>s of the body corresp<strong>on</strong>ding to different emoti<strong>on</strong>s [5-7]. Examples of those activati<strong>on</strong>s<strong>and</strong> inactivati<strong>on</strong> are paralysis of muscles in case of fear, increase of heart rate <strong>and</strong> sudati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g>aroused emoti<strong>on</strong>s. They are generally less perceivable by people, unless an observer is closeenough to the pers<strong>on</strong> that feels the emoti<strong>on</strong>. However these reacti<strong>on</strong>s could be easily recordedusing specific sensors. Some of those physiological changes can also be directly perceived suchas a str<strong>on</strong>g increase of blood pressure that would lead to a blush of the cheeks.All those modalities could be used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> recogniti<strong>on</strong>, but so far, most of the studiesc<strong>on</strong>cerning machine based emoti<strong>on</strong> assessment have used facial expressi<strong>on</strong>s <strong>and</strong>/or speech.However, we believe that physiological signals have several advantages compared to video <strong>and</strong>sound:- the sensors used to record those signals are closer to the body since they are generallydirectly placed <strong>on</strong> the user, reducing potential sources of noise <strong>and</strong> problems due to theunavailability of the signal (user not turning the head in fr<strong>on</strong>t of the camera or notspeaking);- they have very good time resp<strong>on</strong>ses, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance muscle activity can be detected earlierby using electromyography (EMG) than by using a camera;- they are harder to c<strong>on</strong>trol, it is thus harder to fake an emoti<strong>on</strong>;- in the case of impaired users that cannot move facial muscles or express themselves,many physiological signals, such as brain waves, are still usable <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment.Since emoti<strong>on</strong>s are clearly multimodal processes it is necessary to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m fusi<strong>on</strong> of differentmodalities to reliably assess them. This could be d<strong>on</strong>e by fusing different physiological signalsbut also by fusi<strong>on</strong> with other modalities such as postures, gestures <strong>and</strong> speech. Fusing thosedifferent sources could help to increase the accuracy of emoti<strong>on</strong> assessment; it is alsoadvantageous when some signals are not available because of technical problems like3


Chapter 1disc<strong>on</strong>nected or damaged sensors. Moreover, the synchr<strong>on</strong>izati<strong>on</strong> between the activati<strong>on</strong> ofdifferent modalities could also be a potential source of emoti<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>.The goal of this thesis is to investigate the usability of physiological modalities to improveaffective computing methods with the l<strong>on</strong>g-term goal of enhancing HMI. For this purpose, thiswork will focus <strong>on</strong> the assessment of a user’s emoti<strong>on</strong>al states from physiological signals ofdifferent nature. More specifically we are interested in the use of brain signals in c<strong>on</strong>juncti<strong>on</strong>with more classical physiological signals as will be detailed in Chapter 2.1.2.2 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment as a comp<strong>on</strong>ent of HCIFigure 1.2 presents a framework describing how emoti<strong>on</strong> assessment could be integrated as acomp<strong>on</strong>ent of Human-Computer Interacti<strong>on</strong> (HCI). As proposed by Norman [8] the interacti<strong>on</strong>with a machine, from the point of view of the user, can be decomposed in executi<strong>on</strong> / evaluati<strong>on</strong>cycles. After identifying his / her goals, the user starts an executi<strong>on</strong> stage. It c<strong>on</strong>sists in<str<strong>on</strong>g>for</str<strong>on</strong>g>mulating his / her intenti<strong>on</strong>s, specifying the necessary sequence of acti<strong>on</strong>s <strong>and</strong> executing thoseacti<strong>on</strong>s. Next, the computer executes the given comm<strong>and</strong>s <strong>and</strong> output results through availablemodalities. The sec<strong>on</strong>d stage is the evaluati<strong>on</strong> which is realized by: perceiving computer outputs,interpreting them <strong>and</strong> evaluating the outcome (i.e. are the goals satisfied?).MachineComm<strong>and</strong> executi<strong>on</strong><str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al adaptati<strong>on</strong>Decisi<strong>on</strong> of proper reacti<strong>on</strong><str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> synthesis<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessmentSignal acquisiti<strong>on</strong>Feature extracti<strong>on</strong>Classificati<strong>on</strong>Executi<strong>on</strong>Intenti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g>mulati<strong>on</strong>Acti<strong>on</strong>s specificati<strong>on</strong>Executi<strong>on</strong> of acti<strong>on</strong>sEvaluati<strong>on</strong>Percepti<strong>on</strong>Interpretati<strong>on</strong>Outcome evaluati<strong>on</strong><str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al evaluati<strong>on</strong> / elicitati<strong>on</strong><str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>alcuesUserGoalsFigure 1.2. <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment in human computer interfaces, adapted from the executi<strong>on</strong> / evaluati<strong>on</strong> model[8].4


Introducti<strong>on</strong>According to the cognitive theory of emoti<strong>on</strong>s, emoti<strong>on</strong>s are issued from a cognitive processcalled appraisal that evaluates a stimulus according to several criteria such as goal relevance <strong>and</strong>c<strong>on</strong>sequences of the event [9-11]. For this reas<strong>on</strong>, an emoti<strong>on</strong>al evaluati<strong>on</strong> step, corresp<strong>on</strong>ding tothe appraisal process, was added in Figure 1.2 at the evaluati<strong>on</strong> stage. Elicitati<strong>on</strong> of emoti<strong>on</strong>s isknown to be related to changes in several comp<strong>on</strong>ents of the organism such as physiological,motor <strong>and</strong> behavioral comp<strong>on</strong>ents [10]. It is thus possible to c<strong>on</strong>sider those changes as emoti<strong>on</strong>alcues that can be used to automatically detect the elicited emoti<strong>on</strong> after being recorded by theadequate sensors. The detected emoti<strong>on</strong> can then be used to adapt the interacti<strong>on</strong> by modifyingcomm<strong>and</strong> executi<strong>on</strong>. The in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> presented <strong>on</strong> the output modalities can also be influenceddirectly by the emoti<strong>on</strong>al adaptati<strong>on</strong>, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance by synthesizing an emoti<strong>on</strong>al resp<strong>on</strong>se <strong>on</strong>screens <strong>and</strong> speakers.This thesis will focus <strong>on</strong> the green parts of Figure 1.2 to evaluate the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance ofphysiological signals <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment. For this purpose physiological signals should beacquired during emoti<strong>on</strong>al stimulati<strong>on</strong>. This is often d<strong>on</strong>e by designing protocols to elicitemoti<strong>on</strong>s. Three protocols were designed during this thesis with <strong>on</strong>e of them corresp<strong>on</strong>ding to aHCI c<strong>on</strong>text. Once the signals are acquired, some of their features that are known to be related toemoti<strong>on</strong>al reacti<strong>on</strong>s are extracted. The computed features are then used to train a classifier thatfinds a computati<strong>on</strong>al model mapping physiological features values to a given emoti<strong>on</strong>al state.This model (classifier) can then be used to recover the emoti<strong>on</strong>al state corresp<strong>on</strong>ding to a newinstance of the features.Several applicati<strong>on</strong>s can be derived from the presented framework, some of them going bey<strong>on</strong>dhuman-computer interfaces to reach human-machine interfaces in general <strong>and</strong> even humanhumaninterfaces.1.2.3 Applicati<strong>on</strong>s of emoti<strong>on</strong> assessmentWhen thinking about emoti<strong>on</strong> recogniti<strong>on</strong> <strong>on</strong>e applicati<strong>on</strong> that generally comes to mind is the liedetector. This is certainly due to the fact that it was the first tool able to estimate the true internalstate of some<strong>on</strong>e without him / her wanting it, which led to the debates that we all know.However, this secti<strong>on</strong> will show that the lie detector is just the top of the iceberg <strong>and</strong> that manymore applicati<strong>on</strong>s can be targeted by research <strong>on</strong> emoti<strong>on</strong> assessment. We believe that most ofthem are more ethical <strong>and</strong> less subject to debate, essentially because their goal is not to revealin<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> that could be seen as private, but rather to improve the way machines react to humanfeelings, prevent accidents <strong>and</strong> rehabilitate impaired people. The following list of applicati<strong>on</strong>s isof course not exhaustive but is rather made to give some insights into the possibilities <strong>and</strong>advantages of emoti<strong>on</strong> assessment technologies as well as to provide a better underst<strong>and</strong>ing ofthe c<strong>on</strong>text <strong>and</strong> stakes of this thesis.5


Chapter 1a. Human machine interfaces<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment can be used to automatically evaluate the user experience of software. Userexperience design aims at integrating the percepti<strong>on</strong> of the system by the user in the developmentof applicati<strong>on</strong>s <strong>and</strong> devices. This mainly c<strong>on</strong>sists of asking users to fill in surveys about theiremoti<strong>on</strong>al state <strong>and</strong> m<strong>on</strong>itoring their facial expressi<strong>on</strong>s <strong>and</strong> physiological signals while theymanipulate different interfaces [12]. In this case, offline automatic assessment of emoti<strong>on</strong>s helpsto avoid the manual <strong>and</strong> visual processing of large amount of gathered data. But certainly <strong>on</strong>e ofthe main interests <str<strong>on</strong>g>for</str<strong>on</strong>g> HMI is to adjust <strong>on</strong>line the behavior of the machine to the emoti<strong>on</strong>al stateof the user (i.e. directly after or during the feeling of the emoti<strong>on</strong>).As stated in Secti<strong>on</strong> 1.1, the inclusi<strong>on</strong> of emoti<strong>on</strong>s in the human-machine loop can help toimprove <strong>and</strong> disambiguate communicati<strong>on</strong>. For instance, when the user repeats a comm<strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g>the sec<strong>on</strong>d time <strong>and</strong> agitati<strong>on</strong> or stress is detected as the current emoti<strong>on</strong>, the system could inferthat it should try to change its previous resp<strong>on</strong>se because the result was not satisfying. There areat least two particular applicati<strong>on</strong>s that could be cited to illustrate adaptati<strong>on</strong> of HMI to humanemoti<strong>on</strong>s: learning plat<str<strong>on</strong>g>for</str<strong>on</strong>g>ms <strong>and</strong> games.In a learning situati<strong>on</strong> many affective states can be elicited. Discouragement, anxiety, excitati<strong>on</strong>,curiosity, are just some examples of the whole space of possible emoti<strong>on</strong>s. Some of thoseemoti<strong>on</strong>s are favorable to knowledge acquisiti<strong>on</strong> (i.e. curiosity) while some others are not (i.e.discouragement), some could also be useful as l<strong>on</strong>g as their intensity does not exceed a certainthreshold (i.e. anxiety <strong>and</strong> excitati<strong>on</strong>). With the development of e-learning, serious games <strong>and</strong>other learning plat<str<strong>on</strong>g>for</str<strong>on</strong>g>ms [13-15], it is now more <strong>and</strong> more comm<strong>on</strong> to learn through the use ofcomputer interfaces. In a st<strong>and</strong>ard class-room the teacher is generally able to react according tothe feeling of participants but this is not the case with st<strong>and</strong>ard computer interfaces. In thissituati<strong>on</strong>, automatically adapting the learning strategy of the system to the emoti<strong>on</strong>al state of theuser could foster the acquisiti<strong>on</strong> of knowledge. Adaptati<strong>on</strong> could be d<strong>on</strong>e trough the regulati<strong>on</strong> ofst<strong>and</strong>ards criteria like the speed of the in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> flow, the difficulty of exercises <strong>and</strong> questi<strong>on</strong>sbut it could also be d<strong>on</strong>e by providing advices <strong>and</strong> tips to the learner. Another possibility is tohave a virtual agent that could show artificial empathy by providing encouragements <strong>and</strong>displaying appropriate emoti<strong>on</strong>al cues [16]. In all those cases, the idea is to try to motivate thelearner toward a state of mind that will increase his / her per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance <strong>and</strong> learning outcomes.Games are also interesting from a HMI point of view because they are an ideal ground <str<strong>on</strong>g>for</str<strong>on</strong>g> thedesign of new ways to communicate with the machine. One of the main goals of games is toprovide emoti<strong>on</strong>al experiences such as fun <strong>and</strong> excitement which generally occurs when theplayer is str<strong>on</strong>gly involved in the course of the game acti<strong>on</strong>. However, a loss of involvement canoccur if the game does not corresp<strong>on</strong>d to the player’s expectati<strong>on</strong>s <strong>and</strong> competences; he might6


Introducti<strong>on</strong>then feel emoti<strong>on</strong>s like boredom, distress <strong>and</strong> frustrati<strong>on</strong> [17, 18]. In this situati<strong>on</strong>, the c<strong>on</strong>tent ofthe game could be adapted to better fit the player <strong>and</strong> game designer’s expectati<strong>on</strong>s. For instance,in a role playing game, if the player seems to have more fun by exchanging in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> withn<strong>on</strong>-players characters <strong>and</strong> solving puzzles than fighting m<strong>on</strong>sters, the game could automaticallyincrease this type of events. Some games are also purposely inducing particular emoti<strong>on</strong>al statessuch as fear in the case of a horror game. <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> detecti<strong>on</strong> could then be used to c<strong>on</strong>trol theeffectiveness of emoti<strong>on</strong> elicitati<strong>on</strong> <strong>and</strong> adapt the elicitati<strong>on</strong> methods <str<strong>on</strong>g>for</str<strong>on</strong>g> better success.<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al states could also be used to modulate movements of the player character such ashaving a worse precisi<strong>on</strong> at shooting in case of stress <strong>and</strong> limited movements in the case of fearto simulate paralysis. Finally, another possibility is to adapt the difficulty of the game accordingto the felt emoti<strong>on</strong>. This can avoid falling in the extreme cases of too easy <strong>and</strong> too hard gamesthat would respectively elicit boredom <strong>and</strong> frustrati<strong>on</strong>. This later approach will be discussed inChapter 7.b. Behavior predicti<strong>on</strong> <strong>and</strong> m<strong>on</strong>itoring of critical states<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s are known to modulate tendencies to acti<strong>on</strong> [10, 19, 20]. For instance when some<strong>on</strong>e isfeeling a positive emoti<strong>on</strong> he / she will rather tend to approach the stimuli, while fear willgenerally result in withdrawal from the situati<strong>on</strong>. <str<strong>on</strong>g>Based</str<strong>on</strong>g> <strong>on</strong> this fact, emoti<strong>on</strong> detecti<strong>on</strong> is anindicator <str<strong>on</strong>g>for</str<strong>on</strong>g> behavior predicti<strong>on</strong>. It is thus possible to m<strong>on</strong>itor critical emoti<strong>on</strong>al states that couldlead to potential harmful or dangerous behaviors.Examples of such applicati<strong>on</strong>s could be to m<strong>on</strong>itor stress while driving to avoid accidents byencouraging the driver to stop <strong>and</strong> calm down. On a similar basis, some operators have tosupervise critical operati<strong>on</strong>s <strong>and</strong> a loss of task engagement could induce severe damage.M<strong>on</strong>itoring of this cognitive state could then be of interest to re-engage the supervisor in his task.<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s can also be m<strong>on</strong>itored while per<str<strong>on</strong>g>for</str<strong>on</strong>g>ming dangerous operati<strong>on</strong>s, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance to <str<strong>on</strong>g>for</str<strong>on</strong>g>bidthe use of potentially harmful comm<strong>and</strong>s in the case of too high stress. Finally, m<strong>on</strong>itoring ofemoti<strong>on</strong>al states could also provide in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> to determine when an elderly or disabled pers<strong>on</strong>needs help, without having that pers<strong>on</strong> per<str<strong>on</strong>g>for</str<strong>on</strong>g>m usual acti<strong>on</strong>s like giving a ph<strong>on</strong>e call to ask <str<strong>on</strong>g>for</str<strong>on</strong>g>help.c. In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> indexing <strong>and</strong> retrievalThe quantity of multimedia data (images, videos, sounds, music …) that is available <strong>on</strong> theinternet is increasing. Moreover, availability of easy to use multimedia recorders embedded inlow cost mobile devices leads to an incredible amount of self-collected data. This will be furtherencouraged by the coming of ubiquitous computing: the embedding of in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> processingtools in objects [21]. Recording would then be possible at any moment <strong>and</strong> any place.7


Chapter 1For this reas<strong>on</strong>, there is a real need of new technology to index <strong>and</strong> intelligently retrievein<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>. According to Hanjalic [22] retrieving in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> could be d<strong>on</strong>e by eitherhighlighting particular items of interest or by summarizing the c<strong>on</strong>tent of a complete recording.This could be d<strong>on</strong>e by finding in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> that c<strong>on</strong>tains str<strong>on</strong>g emoti<strong>on</strong>al c<strong>on</strong>tent [23, 24]. In thisc<strong>on</strong>text emoti<strong>on</strong> detecti<strong>on</strong> could be used to c<strong>on</strong>struct individual profiles that would store therelati<strong>on</strong>s between multimedia features <strong>and</strong> emoti<strong>on</strong>al states. In this way the search of emoti<strong>on</strong>alc<strong>on</strong>tent would be adapted to each pers<strong>on</strong> according to his / her profile <strong>and</strong> his / her subjectiveinterpretati<strong>on</strong> of multimedia features.Events that people would like to record generally coincide with str<strong>on</strong>g emoti<strong>on</strong>al states likepleasure <strong>and</strong> joy. Thus, emoti<strong>on</strong> detecti<strong>on</strong> could be used, in a ubiquitous envir<strong>on</strong>ment, to detectthose events <strong>and</strong> automatically trigger the recording of relevant scene of the every-day life. At thesame time, this data could be indexed emoti<strong>on</strong>ally <strong>and</strong> retrieved later by using emoti<strong>on</strong>al tags.Automatic emoti<strong>on</strong>al tagging could also be per<str<strong>on</strong>g>for</str<strong>on</strong>g>med at a later stage by presenting previouslyrecorded data to participants while m<strong>on</strong>itoring their emoti<strong>on</strong>al states. This last point is an activetopic of research being c<strong>on</strong>ducted within the European Network of Excellence PetaMedia 1 towhich we participate.d. Health <strong>and</strong> rehabilitati<strong>on</strong> applicati<strong>on</strong>sEmpirical data analysis has shown that emoti<strong>on</strong>s are factors that can influence the speed <strong>and</strong>efficiency of treatments that are provided during a cure [1]. Thus, emoti<strong>on</strong> assessment can beused as any other m<strong>on</strong>itoring device to c<strong>on</strong>trol patient’s signs of rapid <strong>and</strong> effective recovery.However its usability <strong>and</strong> effectiveness would certainly be more important to help <str<strong>on</strong>g>for</str<strong>on</strong>g> the healingof disorders implying emoti<strong>on</strong>al <strong>and</strong> social impairments such as depressi<strong>on</strong> <strong>and</strong> autism. Anexample <str<strong>on</strong>g>for</str<strong>on</strong>g> autism is given by [25] where an emoti<strong>on</strong>al robot is designed to play with autisticchildren. The idea behind the use of a robot is that autistic children would prefer to interact with amachine that has simplified rules of communicati<strong>on</strong> <strong>and</strong> is thus less intimidating. In thisapplicati<strong>on</strong>, emoti<strong>on</strong> assessment is used to adapt the behavior of the robot (speed <strong>and</strong> directi<strong>on</strong> ofa basketball basket, background music <strong>and</strong> game objective) according to the preferences of thechild. The objective is to have a robot that is able to change its behavior al<strong>on</strong>g time like a realtherapist would do.Some accidents <strong>and</strong> diseases, like amyotrophic lateral sclerosis, have as a c<strong>on</strong>sequence thecomplete paralysis of the body. Disabled pers<strong>on</strong>s that suffer from this type of trauma are not ableto communicate through st<strong>and</strong>ard channels since this generally requires the activati<strong>on</strong> of severalmuscles. In order to give them the possibility to communicate, <strong>on</strong>e soluti<strong>on</strong> is to use interfaces1 http://www.petamedia.eu/ (Retrieved <strong>on</strong> 30 April 2009)8


Introducti<strong>on</strong>that do not rely <strong>on</strong> st<strong>and</strong>ard peripheral nerves <strong>and</strong> muscles but directly <strong>on</strong> the brain neur<strong>on</strong>alactivati<strong>on</strong>s. Such interfaces are named <strong>Brain</strong>-Computer Interfaces (BCI) [26-28]. Current BCIaim to detect brain activity that corresp<strong>on</strong>ds to complex tasks (mental calculus, imaginati<strong>on</strong> offinger taping, etc.) that are traduced in comm<strong>and</strong>s like moving a mouse cursor <strong>and</strong> choosing aletter from the alphabet [27, 29, 30]. Generally the user needs training be<str<strong>on</strong>g>for</str<strong>on</strong>g>e using such systems.In case the objective of the user is to express an emoti<strong>on</strong>, classical BCI tasks (e.g., imaginati<strong>on</strong> offinger tapping) seem to be really far from this objective <strong>and</strong> it is more appropriate to use mentaltasks such as the remembering of a similar emoti<strong>on</strong>al event (see Chapter 6). This emoti<strong>on</strong>elicitati<strong>on</strong> task can then be regarded as a mental task that the user tries to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m in order tocommunicate his / her feelings. The objective is then to detect the resulting emoti<strong>on</strong>al states frombrain signals <strong>and</strong> output it in order to give back to disabled people the ability to express theirfeelings [31].1.2.4 Related questi<strong>on</strong>s / problems<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s have been studied from several perspectives <strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> many centuries starting withdiscussi<strong>on</strong>s of philosophers such as Aristotle [1, 32]. Darwin was also a pi<strong>on</strong>eer who tried toexplain the origin of emoti<strong>on</strong>s <strong>and</strong> their importance <str<strong>on</strong>g>for</str<strong>on</strong>g> the survival of species [33, 34]. But it is<strong>on</strong>ly recently (1950’s - 1960’s), with the emergence of cognitive sciences in psychology, thatemoti<strong>on</strong>s gathered a real interest. Thus the study of emoti<strong>on</strong>s in psychology is rather a new fieldof research <strong>and</strong> there are still a lot of issues that remain unsolved. Those issues have a str<strong>on</strong>gimportance <str<strong>on</strong>g>for</str<strong>on</strong>g> HMI researchers since the definiti<strong>on</strong>, elicitati<strong>on</strong> <strong>and</strong> characterizati<strong>on</strong> of emoti<strong>on</strong>sis the basis <str<strong>on</strong>g>for</str<strong>on</strong>g> proper emoti<strong>on</strong> assessment, synthesis <strong>and</strong> adaptati<strong>on</strong>. This secti<strong>on</strong> will emphasizethe issues that are most relevant to emoti<strong>on</strong> assessment.When designing a system <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment <strong>on</strong>e of the critical steps is to specify theemoti<strong>on</strong>s that will be detected. This questi<strong>on</strong> is directly related to the issue of the definiti<strong>on</strong> ofemoti<strong>on</strong>s: what is an emoti<strong>on</strong>, how many emoti<strong>on</strong>s exist, are there emoti<strong>on</strong>s more important thanothers? The representati<strong>on</strong> <strong>and</strong> modeling of emoti<strong>on</strong>s is a subject of debate in psychology. As ac<strong>on</strong>sequence several models that account <str<strong>on</strong>g>for</str<strong>on</strong>g> different comp<strong>on</strong>ents of emoti<strong>on</strong>s were developed.The representati<strong>on</strong> <strong>and</strong> models of emoti<strong>on</strong>s that were found to be of interest in the specific caseof emoti<strong>on</strong> assessment will be presented in Chapter 2. The easiest way to represent an emoti<strong>on</strong> iscertainly by using words like fear, anger <strong>and</strong> joy. However, there is an incredible amount ofvocabulary to refer to our affective states. In the dicti<strong>on</strong>ary of affect in language Whissel [35]registered up to 3000 English words with affective c<strong>on</strong>notati<strong>on</strong>s. Those words do not always havethe same meaning from <strong>on</strong>e individual to another <strong>and</strong> it is sometimes difficult to determine if aword refers to an emoti<strong>on</strong> or not (think to “disobedient” <str<strong>on</strong>g>for</str<strong>on</strong>g> instance that is part of Whisseldicti<strong>on</strong>ary). Words are also culture <strong>and</strong> language dependent <strong>and</strong> the difference between twowords is sometimes difficult to catch such as <str<strong>on</strong>g>for</str<strong>on</strong>g> joy <strong>and</strong> happiness. From an HMI point of view,9


Chapter 1the questi<strong>on</strong> of the relative importance of some emoti<strong>on</strong> labels to others is str<strong>on</strong>gly dependent <strong>on</strong>the applicati<strong>on</strong>. For instance, detecting disgust is not really interesting to adapt the learningstrategy of a virtual teacher but it could be very useful to index horror movies. It is thus reallydifficult to choose the appropriate representati<strong>on</strong> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> the general purpose of emoti<strong>on</strong>assessment while there is a str<strong>on</strong>g importance of designing <strong>and</strong> choosing emoti<strong>on</strong> models that arec<strong>on</strong>sistent with the targeted applicati<strong>on</strong>.Another point of interest is the questi<strong>on</strong> of subjectivity that can occur at several levels. If twopers<strong>on</strong>s are stimulated by the same event, they can feel very different emoti<strong>on</strong>s depending <strong>on</strong>several criteria such as their past experience <strong>and</strong> their current goals. For instance if a pers<strong>on</strong> ishungry, presenting him or her with chocolate can elicit pleasure, while it would elicit disgust inthe case the pers<strong>on</strong> would be sick of chocolate. It is thus really difficult to infer the emoti<strong>on</strong>alstate of some<strong>on</strong>e just by having in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> about the event that elicited the emoti<strong>on</strong>. Sincedrawing a complete profile of some<strong>on</strong>e goals <strong>and</strong> past experiences is quite unfeasible, usingemoti<strong>on</strong>al outputs is a promising soluti<strong>on</strong> to automatically infer emoti<strong>on</strong>s. Associating a stimulusto an emoti<strong>on</strong>al label is thus difficult, raising the questi<strong>on</strong> of the c<strong>on</strong>structi<strong>on</strong> of emoti<strong>on</strong>aldatabases that need to associate recorded data with a specific emoti<strong>on</strong>. A soluti<strong>on</strong> is to ask theuser to report about his / her feelings exactly as if an expert was asked to give labels to objects ofinterest. This supposes that a pers<strong>on</strong> is an expert in evaluating his / her feelings. But is it reallythe case? Not always, according to psychologists, since they separate the objectively felt emoti<strong>on</strong>(change of the organism state) from the resulting subjective feeling (the self-perceived emoti<strong>on</strong>).All those remarks have str<strong>on</strong>g implicati<strong>on</strong>s <strong>on</strong> the design of protocols to gather emoti<strong>on</strong>al data<strong>and</strong> raise several questi<strong>on</strong>s: how to elicit an emoti<strong>on</strong> with maximum certainty <strong>and</strong> how toannotate the collected data? There are no clear answers to those questi<strong>on</strong>s <strong>and</strong> assumpti<strong>on</strong>s willhave to be made <str<strong>on</strong>g>for</str<strong>on</strong>g> the recording of emoti<strong>on</strong>al collecti<strong>on</strong>s.The use of physiological signals also raises a lot of questi<strong>on</strong>s. What type of physiological signalsshould be used? Is the physiological resp<strong>on</strong>se to emoti<strong>on</strong>al stimuli always the same? There aremany physiological signals that have been shown to correlate with emoti<strong>on</strong>s. Chapter 2 discussesmost of these signals. However, it has been shown that there are differences in their activati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g>a given emoti<strong>on</strong> elicited in different c<strong>on</strong>texts. Thus an emoti<strong>on</strong> detecti<strong>on</strong> system should bedesigned <str<strong>on</strong>g>for</str<strong>on</strong>g> a particular applicati<strong>on</strong> or take into account c<strong>on</strong>textual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>. Simultaneouslymeasuring various physiological signals requires the use of several sensors that are sometimesquite obtrusive since they can m<strong>on</strong>opolize the use of <strong>on</strong>e h<strong>and</strong> <strong>and</strong> are not com<str<strong>on</strong>g>for</str<strong>on</strong>g>table. The priceof those sensors should also to be taken into account. For these reas<strong>on</strong>s the issue of determiningthe most useful sensors is of importance. Finally, there is also variability in physiological signals,from pers<strong>on</strong> to pers<strong>on</strong> but also from day to day, that yield to difficulty in designing a system that10


Introducti<strong>on</strong>will detect emoti<strong>on</strong>s accurately <str<strong>on</strong>g>for</str<strong>on</strong>g> every<strong>on</strong>e <strong>and</strong> everyday. Methods to cope with this variabilityhave to be developed.1.3 C<strong>on</strong>tributi<strong>on</strong>sThe main objective of this thesis is to evaluate the usability of physiological signals as emoti<strong>on</strong>alcues <str<strong>on</strong>g>for</str<strong>on</strong>g> assessing emoti<strong>on</strong>s in the c<strong>on</strong>text of affective computing. More precisely, per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance isanalyzed <str<strong>on</strong>g>for</str<strong>on</strong>g> signals from the central nervous system <strong>and</strong> the peripheral nervous systemseparately as well as <str<strong>on</strong>g>for</str<strong>on</strong>g> the fusi<strong>on</strong> of these two types of physiological measures.The c<strong>on</strong>tributi<strong>on</strong>s of this thesis are: Identificati<strong>on</strong> of key points in computerized emoti<strong>on</strong> assessment: issues such as thedefiniti<strong>on</strong> of emoti<strong>on</strong>al classes of interest, the choice of a ground truth <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>classificati<strong>on</strong> <strong>and</strong> the durati<strong>on</strong> of emoti<strong>on</strong>al epochs are investigated. Setting up acquisiti<strong>on</strong> protocols <str<strong>on</strong>g>for</str<strong>on</strong>g> different elicitati<strong>on</strong> c<strong>on</strong>texts: three different emoti<strong>on</strong>alstimuli were employed, namely images, recall of past emoti<strong>on</strong>al events <strong>and</strong> playing agame at several difficulty levels. Evaluati<strong>on</strong> of classificati<strong>on</strong> algorithms <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment: comparis<strong>on</strong> of differentclassificati<strong>on</strong> (Naive-Bayes, LDA, SVM, RVM), feature selecti<strong>on</strong> (ANOVA, FCBF,SFFS, Fisher) <strong>and</strong> fusi<strong>on</strong> (feature <strong>and</strong> classifier level) techniques was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med. Dem<strong>on</strong>strati<strong>on</strong> of usefulness of EEGs (ElectroEncephaloGrams) <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment:the accuracy of emoti<strong>on</strong> assessment from EEG features was compared to the accuracyobtained with peripheral features <strong>on</strong> different emoti<strong>on</strong>al classes <strong>and</strong> c<strong>on</strong>texts. It wasshown that EEG features can per<str<strong>on</strong>g>for</str<strong>on</strong>g>m better than peripheral features especially <str<strong>on</strong>g>for</str<strong>on</strong>g> shorttermassessment. Dem<strong>on</strong>strati<strong>on</strong> of the usefulness of fusi<strong>on</strong> between peripheral <strong>and</strong> EEG modalities: thefusi<strong>on</strong> was d<strong>on</strong>e at different levels (features <strong>and</strong> decisi<strong>on</strong> levels) with results showing theinterest of fusi<strong>on</strong>.Those c<strong>on</strong>tributi<strong>on</strong>s were included in a number of publicati<strong>on</strong>s in internati<strong>on</strong>al c<strong>on</strong>ferences [36-38] <strong>and</strong> a journal [39]. These publicati<strong>on</strong>s are also listed in Appendix E together with otherc<strong>on</strong>tributi<strong>on</strong>s <strong>and</strong> collaborati<strong>on</strong>s related to the study of physiological signals <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>assessment.1.4 Thesis overviewThis thesis is divided in 8 chapters.11


Chapter 1Chapter 2 presents a state of the art regrouping the important noti<strong>on</strong>s c<strong>on</strong>cerning emoti<strong>on</strong>assessment from physiological signals. It addresses several models <strong>and</strong> definiti<strong>on</strong>s of emoti<strong>on</strong>sproposed in the fields of psycho-physiology <strong>and</strong> sociology. The different physiological signalsrelated to emoti<strong>on</strong>al processes are reviewed together with the existing methods <strong>and</strong> sensorsusable <str<strong>on</strong>g>for</str<strong>on</strong>g> the acquisiti<strong>on</strong> of those signals. Finally, the results of studies using physiologicalsignals to assess emoti<strong>on</strong>s are reported <strong>and</strong> discussed according to several criteria.Chapter 3 c<strong>on</strong>cerns the acquisiti<strong>on</strong> of physiological signals. It first presents the material used <str<strong>on</strong>g>for</str<strong>on</strong>g>the acquisiti<strong>on</strong> of physiological signals. It also provides some recommendati<strong>on</strong>s regarding how topositi<strong>on</strong> the sensors <strong>and</strong> c<strong>on</strong>trol the quality of the signals. The features extracted from the signalsto characterize the physiological activity are then presented.Chapter 4 describes the methods employed to assess emoti<strong>on</strong>s from the extracted features. It firstaddresses the possible methods available to c<strong>on</strong>struct a ground-truth, i.e. the true emoti<strong>on</strong>al stateassociated to a physiological activity. In the sec<strong>on</strong>d part the algorithms used <str<strong>on</strong>g>for</str<strong>on</strong>g> the classificati<strong>on</strong>of those emoti<strong>on</strong>al states are presented together with the measure chosen to compare theirper<str<strong>on</strong>g>for</str<strong>on</strong>g>mances. Finally, the feature selecti<strong>on</strong> methods used to reduce the amount of features aredetailed.Chapter 5, 6 <strong>and</strong> 7 present the results of emoti<strong>on</strong> assessment from physiological signals <str<strong>on</strong>g>for</str<strong>on</strong>g>different emoti<strong>on</strong> elicitati<strong>on</strong> strategies <strong>and</strong> classificati<strong>on</strong> methods. In Chapter 5, the emoti<strong>on</strong>swere elicited by using pictures <strong>and</strong> different ground-truths were compared. In Chapter 6, a selfinducti<strong>on</strong>method was employed to elicit emoti<strong>on</strong>s bel<strong>on</strong>ging to three categories (pleasant,unpleasant <strong>and</strong> calm). The per<str<strong>on</strong>g>for</str<strong>on</strong>g>mances of different physiological signals were compared <strong>and</strong>the advantages of fusi<strong>on</strong> of these signals were investigated. In Chapter 7 a gaming paradigm wasproposed to get closer to real HCI applicati<strong>on</strong>s. The per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of emoti<strong>on</strong> assessment fromdifferent signals was investigated <str<strong>on</strong>g>for</str<strong>on</strong>g> several time scales. The changes in physiological activityfollowing game-over events were also analyzed.Chapter 8 presents the c<strong>on</strong>clusi<strong>on</strong>s of this work <strong>and</strong> provides some suggesti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> future work<strong>and</strong> improvements.12


State of the artChapter 2State of the art2.1 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> representati<strong>on</strong>s <strong>and</strong> modelsBe<str<strong>on</strong>g>for</str<strong>on</strong>g>e designing a system able to represent <strong>and</strong> recognize emoti<strong>on</strong>s, it is first necessary to definewhat an emoti<strong>on</strong> is. At the moment, there is no comm<strong>on</strong> framework that can be used to answerthis questi<strong>on</strong>. This is mainly due to the fact that emoti<strong>on</strong>s are complex phenomena that influencemany aspects of everyday life: they help us to avoid danger, take decisi<strong>on</strong>s <strong>and</strong> they also havesome important social values. Over the past century, four main theories of emoti<strong>on</strong>s havedeveloped [33].From a social c<strong>on</strong>structivist point of view, emoti<strong>on</strong>s are <strong>on</strong>ly products of social interacti<strong>on</strong>s <strong>and</strong>cultural rules [33]. This statement is often c<strong>on</strong>sidered too restrictive but nowadays most ofresearchers in affective sciences admit that social <strong>and</strong> cultural rules at least regulate emoti<strong>on</strong>s:<strong>on</strong>e will not have the same emoti<strong>on</strong>al behavior in fr<strong>on</strong>t of <strong>on</strong>e’s senior than in fr<strong>on</strong>t of <strong>on</strong>e’sfriend. On the <strong>on</strong>e h<strong>and</strong>, this view emphasizes the importance that emoti<strong>on</strong>s have in society <strong>and</strong>uncovers applicati<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment in social envir<strong>on</strong>ment, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance in computerizedsocial networks. On the other h<strong>and</strong>, it also suggests that emoti<strong>on</strong>s should be assessed with modelsthat are cultural <strong>and</strong> envir<strong>on</strong>mental specific, making this assessment more complex.Darwinians c<strong>on</strong>sider emoti<strong>on</strong>s as phenomena that are selected by nature according to theirsurvival value, i.e. fear exist because it helps us avoid danger. This is <strong>on</strong>e of the oldest viewsc<strong>on</strong>cerning emoti<strong>on</strong>s <strong>and</strong> has initially been proposed by Darwin in 1872 [33, 34]. The mainimplicati<strong>on</strong> is that emoti<strong>on</strong>s should have identical c<strong>on</strong>structs across individuals <strong>and</strong> thus comm<strong>on</strong>emoti<strong>on</strong>al expressi<strong>on</strong>s <strong>and</strong> behaviors should be found across cultures. This point of view isclearly opposite to social c<strong>on</strong>structivism <strong>and</strong> supports the idea that a single model <str<strong>on</strong>g>for</str<strong>on</strong>g> assessing<strong>and</strong> representing emoti<strong>on</strong>s universally is c<strong>on</strong>ceivable.For Jamesians, emoti<strong>on</strong>s uniquely corresp<strong>on</strong>ds to the feeling of bodily changes, such asmodificati<strong>on</strong>s of heart rate <strong>and</strong> blood pressure, which follow the emoti<strong>on</strong>al stimuli (“I am afraidbecause I shiver”) [33, 40]. Thus if perceptual mechanisms are impaired there is no possibility tofeel emoti<strong>on</strong>s anymore. Although c<strong>on</strong>troversial, this later approach emphasizes the important roleof physiological resp<strong>on</strong>ses in the study of emoti<strong>on</strong>s. More specifically this implies that eachemoti<strong>on</strong> corresp<strong>on</strong>ds to a unique pattern of physiological activity, which is str<strong>on</strong>gly encouragingto go toward emoti<strong>on</strong> assessment from physiological signals.In the cognitive theory, emoti<strong>on</strong>s are issued from a cognitive process called appraisal. Thisprocess is supposed to be “direct <strong>and</strong> n<strong>on</strong>-reflective” <strong>and</strong> evaluates a stimulus according toseveral criteri<strong>on</strong>s such as relevance <strong>and</strong> c<strong>on</strong>sequences of an event [9, 10, 41]. This approachsupports the idea that the brain is the main organ implied in emoti<strong>on</strong>s through organizati<strong>on</strong> of13


State of the artup to thous<strong>and</strong>s of words [35]. What about the remaining emoti<strong>on</strong>s? The psychological point ofview can help to give an answer to that questi<strong>on</strong>.Reference Basic emoti<strong>on</strong>s CriteriaEkman [42]Anger, disgust, fear, joy, Universal facial expressi<strong>on</strong>ssadness, surpriseGray [43] Rage <strong>and</strong> terror, anxiety, joy HardwiredPanksepp [44] Expectancy, fear, rage, panic HardwiredMcDougall [45]Anger, disgust, elati<strong>on</strong>, fear, Relati<strong>on</strong> to instinctssubjecti<strong>on</strong>, tender-emoti<strong>on</strong>,w<strong>on</strong>derMowrer [46] Pain, pleasure Unlearned emoti<strong>on</strong>al statesPlutchik [47]Acceptance, joy, anticipati<strong>on</strong>,anger, disgust, sadness,Relati<strong>on</strong> to adaptive biologicalprocesssurprise, fear.James [40] Fear, grief, love, rage Bodily involvementTable 2.1. Lists of basic emoti<strong>on</strong>s from a biological point of view (from [34]).For several psychologists, an emoti<strong>on</strong> is c<strong>on</strong>sidered as basic if it is an “elementary” <strong>on</strong>e that canbe used to c<strong>on</strong>struct, in combinati<strong>on</strong> with others basic emoti<strong>on</strong>s, a large number (if not unlimited)of n<strong>on</strong> basic emoti<strong>on</strong>s [34]. The word “elementary” signifies that a basic emoti<strong>on</strong> is <strong>on</strong>e thatcannot be decomposed into a combinati<strong>on</strong> of other emoti<strong>on</strong>s. One of the main advantages of thisdefiniti<strong>on</strong> of basic emoti<strong>on</strong>s is that it enables to virtually c<strong>on</strong>struct any emoti<strong>on</strong> as a combinati<strong>on</strong>of basic emoti<strong>on</strong>s like it is possible to mix primary colors to obtain sec<strong>on</strong>dary <strong>on</strong>es. However, theway the combinati<strong>on</strong> is d<strong>on</strong>e differs am<strong>on</strong>g researchers <strong>and</strong> is not really clear. An example of thisview is the wheel of emoti<strong>on</strong>s proposed by Plutchik [48, 49] that is shown in Figure 2.1. Thiswheel is composed of several quadrants c<strong>on</strong>taining the proposed basic emoti<strong>on</strong>s organizedaccording to their closeness: emoti<strong>on</strong>s that are facing each other <strong>on</strong> the wheel are c<strong>on</strong>sidered tobe opposite while emoti<strong>on</strong>s in adjacent quadrants have comm<strong>on</strong> properties (especially if theyhave similar colors). Two adjacent emoti<strong>on</strong>s can then be combined to <str<strong>on</strong>g>for</str<strong>on</strong>g>m a new n<strong>on</strong>-basicemoti<strong>on</strong>. For instance anticipati<strong>on</strong> <strong>and</strong> joy will c<strong>on</strong>stitute optimism. In his wheel, Plutchik alsoadded the c<strong>on</strong>cept of intensity as can be seen from Figure 2.1.Both the psychological <strong>and</strong> the biological views of basic emoti<strong>on</strong>s are challenged by socialc<strong>on</strong>structivism. Since in this theory emoti<strong>on</strong>s are c<strong>on</strong>sidered to be a product of nurture rather thannature, there is no reas<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the existence of basic emoti<strong>on</strong>s in the biological sense. C<strong>on</strong>cerningthe psychological view, social c<strong>on</strong>structivists argue that the choice of the basic emoti<strong>on</strong>s is biasedby cultural influence <strong>and</strong> that the so called “elementary” emoti<strong>on</strong>s are rather emoti<strong>on</strong>s that areprominent in the culture. For this reas<strong>on</strong>, Shaver et al. [50] per<str<strong>on</strong>g>for</str<strong>on</strong>g>med clustering of severalemoti<strong>on</strong>al words using similarity between words evaluated by many participants. The result is ahierarchical tree structure where words are organized in three groups’ levels. The names of the15


Chapter 2groups were chosen as the most comm<strong>on</strong> emoti<strong>on</strong>s or as a close basic emoti<strong>on</strong>. The mainadvantage of this representati<strong>on</strong> is that it allows <str<strong>on</strong>g>for</str<strong>on</strong>g> the tax<strong>on</strong>omy of several emoti<strong>on</strong>al words.Figure 2.1. Plutchik’s wheel of emoti<strong>on</strong>s. (Left) The basic emoti<strong>on</strong> represented as quadrants <strong>and</strong> possiblecombinati<strong>on</strong>s of basic emoti<strong>on</strong>s 2 . (Right) The same wheel with the added c<strong>on</strong>cept of intensity 2 .While representing emoti<strong>on</strong>s with words has the advantage of being instinctive, the questi<strong>on</strong>s ofcultural differences, misunderst<strong>and</strong>ing <strong>and</strong> choice of words are still not resolved. Moreover,labels are discrete <strong>and</strong> hence cannot fully represent some aspects of emoti<strong>on</strong>s like the c<strong>on</strong>tinuousscale of emoti<strong>on</strong>al strength. This is why several researchers focused <strong>on</strong> the search of a c<strong>on</strong>tinuousspace that would represent emoti<strong>on</strong>s with less ambiguity.2.1.2 C<strong>on</strong>tinuous representati<strong>on</strong>sThe idea of searching <str<strong>on</strong>g>for</str<strong>on</strong>g> a c<strong>on</strong>tinuous space able to represent emoti<strong>on</strong>s comes from the cognitivetheory that assumes that any pers<strong>on</strong> possesses an internal representati<strong>on</strong> of emoti<strong>on</strong>s [51]. Thegoal is to find what the dimensi<strong>on</strong>s of this representati<strong>on</strong> are <strong>and</strong> how emoti<strong>on</strong>s are mapped intoit. Several researches (see [51] <str<strong>on</strong>g>for</str<strong>on</strong>g> a review) have generally addressed those questi<strong>on</strong>s byanalyzing participant-reported measures of similarity between either verbal or facial emoti<strong>on</strong>alexpressi<strong>on</strong>s. In those analysis different methods were applied to find a space that minimize (resp.maximize) the distance between similar (resp. different) expressi<strong>on</strong>s. Most of the studies obtaineddifferent spaces but, <strong>on</strong> a closer look, some of the dimensi<strong>on</strong>s were redundant <strong>and</strong> similar acrossstudies.As a result there is nowadays an agreement <strong>on</strong> the first two bipolar <strong>and</strong> most importantdimensi<strong>on</strong>s: valence <strong>and</strong> arousal (Figure 2.2). Valence represents the pleasantness, ranging fromunpleasant to pleasant (it is also called evaluati<strong>on</strong>); while arousal represents the awakeness,2 http://library.thinkquest.org/25500/index2.htm (Retrieved <strong>on</strong> 4 May 2009)16


State of the artalertness <strong>and</strong> activati<strong>on</strong> of the emoti<strong>on</strong> (this dimensi<strong>on</strong> is sometimes called sleep-tensi<strong>on</strong> <strong>and</strong>activati<strong>on</strong>) [20, 52]. There is currently no c<strong>on</strong>sensus <strong>on</strong> a potentially third dimensi<strong>on</strong>. Most of thestudies emphasize that it accounts <str<strong>on</strong>g>for</str<strong>on</strong>g> a low part of the variance in participants judgments,generally leading to its neglect. Otherwise, this third dimensi<strong>on</strong> could be related to c<strong>on</strong>cepts suchas potency, c<strong>on</strong>trol, <strong>and</strong> dominance in order to distinguish between emoti<strong>on</strong>s that are related toapproach <strong>and</strong> withdrawal reacti<strong>on</strong>s.The two-dimensi<strong>on</strong>al valence-arousal space has several advantages. First, it is possible torepresent emoti<strong>on</strong>s without using labels but just a coordinate system that includes emoti<strong>on</strong>almeaning. As a c<strong>on</strong>sequence, any emoti<strong>on</strong> could be represented as a point in this space, even ifthere is no particular label or expressi<strong>on</strong> to define it.Sec<strong>on</strong>dly, since this space was created from the analysis of emoti<strong>on</strong>al expressi<strong>on</strong>s (verbal <strong>and</strong>n<strong>on</strong> verbal), it is possible to associate areas of this space to emoti<strong>on</strong>al labels if necessary (Figure2.2). In [51], the direct mapping of verbal expressi<strong>on</strong>s <strong>on</strong> the space gave the result shown inFigure 2.2.a. As can be seen, labels tend to <str<strong>on</strong>g>for</str<strong>on</strong>g>m a circle within the space. However there aresome evidences that this mapping is not exactly the same from a pers<strong>on</strong> to another. Inc<strong>on</strong>sequence, there are no exact boundaries in the valence-arousal space that define emoti<strong>on</strong>alexpressi<strong>on</strong>s. A soluti<strong>on</strong> could be to represent this variability by defining expressi<strong>on</strong>s as areas thatcan overlap (Figure 2.2.b). Using probabilistic models that define the probability of having agiven expressi<strong>on</strong> knowing the positi<strong>on</strong> in the valence-arousal space could help determine suchboundaries.ArousalArousalFearAngerExcitati<strong>on</strong>FearAngerExcitati<strong>on</strong>Frustrati<strong>on</strong>Happiness / joyFrustrati<strong>on</strong>Happiness / joyValenceValenceSadnessC<strong>on</strong>tentmentSadnessC<strong>on</strong>tentmentBoredomRelaxati<strong>on</strong>BoredomRelaxati<strong>on</strong>Figure 2.2. Valence arousal space with associated labels as (a) points (adapted from [Russell], (adjectives havebeen changed to nouns <strong>and</strong> <strong>on</strong>ly some of the words are displayed <str<strong>on</strong>g>for</str<strong>on</strong>g> clarity) <strong>and</strong> (b) areas.Thirdly, this space can also represent the intensity of the emoti<strong>on</strong> [20, 51]: a point in the middleof the space would represent a low intensive <strong>and</strong> neutral feeling while a point <strong>on</strong> the periphery17


Chapter 2would indicate a full-blown emoti<strong>on</strong>. Finally, it has been shown [53] that the mapping ofemoti<strong>on</strong>al words <strong>on</strong> this space does not vary significantly between four Europeans languages(English, Est<strong>on</strong>ian, Greek <strong>and</strong> Polish). This encourages the use of this space as a way to representemoti<strong>on</strong>s interculturally.As discussed, emoti<strong>on</strong>al labels can be projected from the discrete space of emoti<strong>on</strong>al labels to thec<strong>on</strong>tinuous valence-arousal space. Un<str<strong>on</strong>g>for</str<strong>on</strong>g>tunately there is nothing that guarantees this projecti<strong>on</strong>to be injective so that two different labels could have the same coordinate in the valence-arousalspace. In the representati<strong>on</strong> above (Figure 2.2), this is merely the case <str<strong>on</strong>g>for</str<strong>on</strong>g> fear <strong>and</strong> anger, twohighly negative <strong>and</strong> excited emoti<strong>on</strong>s. In that case the clear distincti<strong>on</strong> between the two emoti<strong>on</strong>scould be achieved by adding the third dimensi<strong>on</strong> of c<strong>on</strong>trol: when <strong>on</strong>e is feeling fear <strong>on</strong>e does nothave c<strong>on</strong>trol over the situati<strong>on</strong> while c<strong>on</strong>trol is present <str<strong>on</strong>g>for</str<strong>on</strong>g> anger.The valence-arousal space has been shown to be effective not <strong>on</strong>ly to represent emoti<strong>on</strong>alexpressi<strong>on</strong>s but also <str<strong>on</strong>g>for</str<strong>on</strong>g> self-assessment of emoti<strong>on</strong>s <strong>and</strong> moods [51, 52, 54, 55]. However, whileRussel [52] argues <str<strong>on</strong>g>for</str<strong>on</strong>g> independent dimensi<strong>on</strong>s of valence <strong>and</strong> arousal, there are some evidencesthat this is not true <str<strong>on</strong>g>for</str<strong>on</strong>g> self-reported feelings. Lang [54] observed that when people assess theirown emoti<strong>on</strong> while watching pictures of the IAPS (Internati<strong>on</strong>al <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Picture System), theirjudgment tends to follow a U-shape centered in the space (Figure 2.3). This shape has also beenobserved using film clips as stimulati<strong>on</strong>s in [24]. This distributi<strong>on</strong> of self-assessments is notsurprising since it is difficult to elicit emoti<strong>on</strong>s that have low arousal <strong>and</strong> high valencesimultaneously as well as emoti<strong>on</strong> of high arousal <strong>and</strong> no valence. This also dem<strong>on</strong>strates thatareas of this space are more important than others <strong>and</strong> that a system per<str<strong>on</strong>g>for</str<strong>on</strong>g>ming emoti<strong>on</strong>assessment should focus <strong>on</strong> them.ArousalValenceFigure 2.3. Self-assessments distributi<strong>on</strong> obtained when eliciting emoti<strong>on</strong>s with images: most of the selfassessedimages lie inside the U-shape.18


State of the art2.1.3 Models of emoti<strong>on</strong>sSecti<strong>on</strong>s 2.1.1 <strong>and</strong> 2.1.2 describe different representati<strong>on</strong>s of emoti<strong>on</strong>s. While they are useful <str<strong>on</strong>g>for</str<strong>on</strong>g>tax<strong>on</strong>omy, categorizati<strong>on</strong> <strong>and</strong> differentiati<strong>on</strong> of emoti<strong>on</strong>s they do not provide much in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>about the process that gives rise to emoti<strong>on</strong>s. Several models, essentially from the cognitive viewof emoti<strong>on</strong>s, have been proposed to answer this questi<strong>on</strong> [9, 10, 32, 56-58]. Most of them arebased <strong>on</strong> the central c<strong>on</strong>cept of appraisal which has been defined by Arnold as the cognitiveprocess that c<strong>on</strong>stantly evaluates the envir<strong>on</strong>ment <strong>and</strong> elicit emoti<strong>on</strong>s [41]. It is generallyc<strong>on</strong>sidered as an unc<strong>on</strong>scious process that is direct <strong>and</strong> n<strong>on</strong>-reflective. To our view, the maindifferences between models of appraisal are the different evaluati<strong>on</strong> criteria they propose.Two of the most famous models are presented in the following secti<strong>on</strong>s: the OCC (Ort<strong>on</strong>y Clore<strong>and</strong> Collins) typology [57], which has been used to define computati<strong>on</strong>al models of emoti<strong>on</strong>s,<strong>and</strong> Scherer’s SECs (Stimulus Evaluati<strong>on</strong> Checks) [9, 10], which can be viewed as an attempt tounify the different emoti<strong>on</strong> theories.a. OCC typologyAccording to Ort<strong>on</strong>y [34], “the defining feature [of an emoti<strong>on</strong>] that we c<strong>on</strong>sider most reas<strong>on</strong>able<strong>and</strong> least c<strong>on</strong>tentious is that the appraisal underlying the emoti<strong>on</strong> be valenced, either positively ornegatively”. It is thus not surprising to find the valence c<strong>on</strong>cept at the heart of his model. In theOCC (Ort<strong>on</strong>y, Clore <strong>and</strong> Collins) typology [57], an emoti<strong>on</strong> is viewed as a valenced reacti<strong>on</strong> to astimulus. More precisely, the elicitati<strong>on</strong> of an emoti<strong>on</strong> relies <strong>on</strong> the evaluati<strong>on</strong> of three maincriteria: the type of the stimulus (agent, object, <strong>and</strong> event), the c<strong>on</strong>cerned entity (self oranother) <strong>and</strong> finally the valence of the emoti<strong>on</strong> (positive or negative). Since each of these criteriacan <strong>on</strong>ly be discretely evaluated, it is possible to represent the model as a tree with the resultingemoti<strong>on</strong>s as the leaves (Figure 2.4).Two examples are given bellow that respectively corresp<strong>on</strong>ds to the green <strong>and</strong> blue dotted line ofFigure 2.4. As a first example, imagine that a student just sat <str<strong>on</strong>g>for</str<strong>on</strong>g> an exam <strong>and</strong> is rather pleased byhis / her per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance. In this case the type of stimuli is an acti<strong>on</strong> (sitting <str<strong>on</strong>g>for</str<strong>on</strong>g> an exam), the agentof interest is the self, <strong>and</strong> since the valence is positive the resulting emoti<strong>on</strong> is pride (green dottedline in Figure 2.4). As a sec<strong>on</strong>d example, our student receives a letter c<strong>on</strong>cerning the result of theexam <strong>and</strong> reads it. Here the stimulus is an event (reading of the letter) <strong>and</strong> the c<strong>on</strong>cerned entity isagain the self. Since prospects are relevant in this case <strong>and</strong> the student has opened the letter, theelicited emoti<strong>on</strong>s will be either satisfacti<strong>on</strong> or disappointment depending of his / her results (bluedotted line in Figure 2.4). Notice that as l<strong>on</strong>g as the letter wasn’t read the elicited emoti<strong>on</strong> washope (he or she was happy about his / her per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance).19


Chapter 2Figure 2.4. The OCC typology (from [57]). Green <strong>and</strong> blue dotted lines corresp<strong>on</strong>d to the examples above.This model has the advantage of being computati<strong>on</strong>ally tractable. C<strong>on</strong>sequently, it has been used<str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of emoti<strong>on</strong> synthesis. Elliot extended it in [59] by adding few emoti<strong>on</strong>s, <strong>and</strong>implemented it in a computati<strong>on</strong>al affective reas<strong>on</strong>er. In [60], the authors c<strong>on</strong>verted this model ina BDI (Belief, Desire <strong>and</strong> Intensi<strong>on</strong>s) architecture <strong>and</strong> also dem<strong>on</strong>strated the effectiveness of thisapproach.This model can also be used <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of user-emoti<strong>on</strong> assessment. In order <str<strong>on</strong>g>for</str<strong>on</strong>g> this modelto be applicable in this case, it is m<strong>and</strong>atory to have a complete knowledge or c<strong>on</strong>trol of theenvir<strong>on</strong>ment in which the user is evolving to be able to follow a path in the tree. This is aninteresting fact since it str<strong>on</strong>gly emphasizes the importance of c<strong>on</strong>text in emoti<strong>on</strong> assessment.However, even with a str<strong>on</strong>g knowledge of c<strong>on</strong>text, valence is still a hard criteri<strong>on</strong> to evaluate.Let’s take the example of a learning applicati<strong>on</strong> since, as explained in the introducti<strong>on</strong>, there areseveral advantages to include emoti<strong>on</strong> assessment in this type of envir<strong>on</strong>ment. Imagine that theuser is in<str<strong>on</strong>g>for</str<strong>on</strong>g>med that he failed in a task then the valence could be guessed as being negative. Now20


State of the artimagine that he just finished the task but he has not been in<str<strong>on</strong>g>for</str<strong>on</strong>g>med about his / her per<str<strong>on</strong>g>for</str<strong>on</strong>g>mancethen the valence will depend <strong>on</strong> the user interpretati<strong>on</strong> of his / her success. Since this in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>is not directly available to the system, it is not possible to infer any emoti<strong>on</strong>al state. In this case,detecting the valence from the measures of user’s emoti<strong>on</strong>al cues can solve the issue.b. SECs theoryIn his theory, Scherer [9, 10] proposed that appraisal can be decomposed into several checks thatare cognitively evaluated. They are called the Stimulus Evaluati<strong>on</strong> Checks (SECs, Table 2.2).Each of these processes is supposed to occur in a specific time span from the elicitati<strong>on</strong> eventwith the possibility of parallel processing (time span that overlap). Those SECs can be grouped inthe four appraisal objectives listed below.- Relevance detecti<strong>on</strong>: the objective of this group of SECs is to determine if the stimulusrequires attenti<strong>on</strong> <strong>and</strong> further processing by analyzing its relevance <str<strong>on</strong>g>for</str<strong>on</strong>g> the pers<strong>on</strong>; thegoal is to answer the questi<strong>on</strong> “Are there any possible implicati<strong>on</strong>s to me or my directenvir<strong>on</strong>ment?”.- Implicati<strong>on</strong> assessment: the next step is then to determine what the implicati<strong>on</strong>s of thestimuli are in terms of c<strong>on</strong>sequences <str<strong>on</strong>g>for</str<strong>on</strong>g> the self. This objective regroups the centralprocesses of appraisal <strong>and</strong> its functi<strong>on</strong> is mainly the protecti<strong>on</strong> <strong>and</strong> the progress of theorganism: “is the stimulus dangerous or appealing?”, “does it corresp<strong>on</strong>d to my goals <strong>and</strong>needs?”. Notice that this appraisal objective is very close to what Darwin defines as thefuncti<strong>on</strong> of emoti<strong>on</strong>s as a whole.- Coping potential determinati<strong>on</strong>: the c<strong>on</strong>cept of coping is well known in psychology <strong>and</strong>is defined as the cognitive mechanisms that are implemented to c<strong>on</strong>trol <strong>and</strong> reduce theimpact of stressful <strong>and</strong> emoti<strong>on</strong>al events. According to Lazarus [32], there are two typesof coping: problem oriented (determinati<strong>on</strong> of an acti<strong>on</strong> to solve the problem that gaverise to the stressful situati<strong>on</strong>) <strong>and</strong> emoti<strong>on</strong> oriented (cognitive regulati<strong>on</strong> of the stress <str<strong>on</strong>g>for</str<strong>on</strong>g>instance by rec<strong>on</strong>sidering the situati<strong>on</strong>). The current appraisal objective is to deal withboth of those aspects.- Normative significance evaluati<strong>on</strong>: the objective of this SECs group is to evaluatewhether the stimulus is compatible with <strong>on</strong>e’s own principles <strong>and</strong> st<strong>and</strong>ards as well aswith social norms <strong>and</strong> values.The order in which the SECs are presented in Table 2.2 is significant as it represents the sequencein which the checks are supposed to be completely evaluated. The appraisal decompositi<strong>on</strong> in thepresent SECs as well as the temporal aspects have been validated in [61, 62].21


Chapter 222AppraisalobjectivesRelevancedetecti<strong>on</strong>Implicati<strong>on</strong>assessmentCopingpotentialdeterminati<strong>on</strong>NormativesignificanceNoveltySECsIntrinsic pleasantnessGoal relevance checkCausal attributi<strong>on</strong>Outcome probabilityDiscrepancy fromexpectati<strong>on</strong>Goal / needc<strong>on</strong>ducivenessUrgencyC<strong>on</strong>trolPowerAdjustmentInternal st<strong>and</strong>ardsExternal st<strong>and</strong>ardsDescripti<strong>on</strong> / Answer the questi<strong>on</strong>Is stimulus novel <strong>and</strong> does it require attenti<strong>on</strong>?Is it familiar <strong>and</strong>/or predictable?Will the stimulus lead to pleasure or pain?This is a property of the stimulus <strong>and</strong> is not related to thecurrent state of the organism (it does not depend <strong>on</strong> thecurrent goal <strong>and</strong> objectives of the subject but could havebeen learned in the past).Does the stimulus have some c<strong>on</strong>sequences <strong>on</strong> mycurrent goals?What or who is the cause of the stimulus <strong>and</strong> why?Often divided in two categories: the resp<strong>on</strong>sible agent<strong>and</strong> the motive.What are the c<strong>on</strong>sequences of the stimulus?What is the probability of each c<strong>on</strong>sequence?To which extent the stimulus is different from what wasexpected.Will the c<strong>on</strong>sequences of the stimulus help me toaccomplish my goals or will they obstruct my goals?Should the stimulus be h<strong>and</strong>led quickly?To which extent the stimulus can be influenced <strong>and</strong>c<strong>on</strong>trolled.If c<strong>on</strong>trol is possible, check the available resources(physical, knowledge, etc.) <str<strong>on</strong>g>for</str<strong>on</strong>g> a potential acti<strong>on</strong>.It is related to the problem oriented coping of Lazarus.If it is not possible to influence the situati<strong>on</strong> (because ofa n<strong>on</strong> c<strong>on</strong>trollable stimulus or lack of resources), checkhow well the organism can adjust with the situati<strong>on</strong>.It is related to the emoti<strong>on</strong> oriented coping of Lazarus.Is the stimulus in accordance with <strong>on</strong>e’s own principles,ideals <strong>and</strong> moral code?Is the stimulus in accordance with social norm <strong>and</strong>values?Table 2.2. List <strong>and</strong> descripti<strong>on</strong> of the different Stimulus Evaluati<strong>on</strong> Checks (SECs) grouped by appraisalobjective <strong>and</strong> temporally ordered.C<strong>on</strong>trarily to the OCC model, where the different criteria are evaluated discretely, Schererproposed that most of the SECs are evaluated <strong>on</strong> c<strong>on</strong>tinuous scales (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance the outcomeprobability SEC is evaluated <str<strong>on</strong>g>for</str<strong>on</strong>g> each event <strong>on</strong> a c<strong>on</strong>tinuous scale ranging from 0 to 1).Moreover, 14 emoti<strong>on</strong>s such as happiness, disgust, anxiety, fear, pride, guilt <strong>and</strong> boredom wereprofiled by giving <str<strong>on</strong>g>for</str<strong>on</strong>g> each of them the possible associated evaluati<strong>on</strong> of the SECs [9] (<str<strong>on</strong>g>for</str<strong>on</strong>g>instance Happiness is associated to high intrinsic pleasantness, medium goal/need relevance, veryhigh outcome probability, very low urgency, etc.). It was shown that these profiles can be used tocorrectly identify an emoti<strong>on</strong> associated with situati<strong>on</strong>s described according to the SECs.


State of the artHowever, the emoti<strong>on</strong>al profiles are described with words like “low”, “high”, “medium” so thatthe computati<strong>on</strong>al implementati<strong>on</strong> of this model could not be directly d<strong>on</strong>e with c<strong>on</strong>tinuousevaluati<strong>on</strong>s of the SECs. A first step should be taken to find evaluati<strong>on</strong> thresholds (<str<strong>on</strong>g>for</str<strong>on</strong>g> each SEC)that define the boundaries between emoti<strong>on</strong>s. Moreover, it still remains that this model is verycomplex <strong>and</strong> requires the evaluati<strong>on</strong> of 13 SECs c<strong>on</strong>firming the remark made <str<strong>on</strong>g>for</str<strong>on</strong>g> the OCC modelabout the difficulty of comp<strong>on</strong>ents evaluati<strong>on</strong>.In his appraisal theory, Scherer does not limit his analysis to the different comp<strong>on</strong>ents of thecognitive appraisal but also studied the relati<strong>on</strong> between the systems (or comp<strong>on</strong>ents) of theorganism taking part in the emoti<strong>on</strong> elicitati<strong>on</strong> <strong>and</strong> differentiati<strong>on</strong> [10]:- the cognitive system, that is resp<strong>on</strong>sible <str<strong>on</strong>g>for</str<strong>on</strong>g> the evaluati<strong>on</strong> of the emoti<strong>on</strong>al event(appraisal of the stimuli);- the aut<strong>on</strong>omic system, that provides support <str<strong>on</strong>g>for</str<strong>on</strong>g> other comp<strong>on</strong>ents (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance thequantity of blood in h<strong>and</strong>s can increase in case of anger to prepare <str<strong>on</strong>g>for</str<strong>on</strong>g> acti<strong>on</strong>);- the motivati<strong>on</strong>al system, related to acti<strong>on</strong> tendencies, urges <strong>and</strong> desires (i.e. will <strong>on</strong>ewithdraw from the stimuli or approach it);- the motor system, to execute the (re)-acti<strong>on</strong> but also to show facial expressi<strong>on</strong>s as well asemoti<strong>on</strong>al gestures <strong>and</strong> behaviors;- the m<strong>on</strong>itor system, which gives rise to the subjective feeling <strong>on</strong>e can experience afterbeing c<strong>on</strong>fr<strong>on</strong>ted to an emoti<strong>on</strong>al situati<strong>on</strong>.This list clearly dem<strong>on</strong>strates the multimodal aspects of emoti<strong>on</strong>s, since emoti<strong>on</strong>al activati<strong>on</strong> isreflected in the activity of all those systems. Notice that this activity could be m<strong>on</strong>itored in orderto assess emoti<strong>on</strong>s; actually it is even possible that m<strong>on</strong>itoring the activity of ALL those systemsis necessary to reliably <strong>and</strong> fully assess emoti<strong>on</strong>s.This last statement goes al<strong>on</strong>g with the comp<strong>on</strong>ential patterning theory. According to this theory,a given evaluati<strong>on</strong> of the SECs leads to a precise activity pattern in the five organism comp<strong>on</strong>ents<strong>and</strong> this pattern can be used to predict emoti<strong>on</strong>. Another important statement is that the differentcomp<strong>on</strong>ents are not independent but rather interrelated. C<strong>on</strong>sequently, they exchange in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong><str<strong>on</strong>g>for</str<strong>on</strong>g> the evaluati<strong>on</strong> of the different SECs. For instance the evaluati<strong>on</strong> of a SEC (cognitivecomp<strong>on</strong>ent) will give rise to activity in another comp<strong>on</strong>ent (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance in the aut<strong>on</strong>omicsystem); this activity could then be feedbacked <str<strong>on</strong>g>for</str<strong>on</strong>g> the evaluati<strong>on</strong> of another SEC. Since the SECsare evaluated in sequence, this theory implies that the pattern of emoti<strong>on</strong>al activity changesduring the appraisal process <strong>and</strong> that synchr<strong>on</strong>izati<strong>on</strong> between the different systems is necessary.23


Chapter 2The collaborati<strong>on</strong> between different comp<strong>on</strong>ents has also been suggested in [63] while thedynamic <strong>and</strong> synchr<strong>on</strong>izati<strong>on</strong> aspects are detailed in [64].2.1.4 Finding an adequate representati<strong>on</strong>In the emoti<strong>on</strong> assessment field, many studies are classifying emoti<strong>on</strong>s into categories such asanger, fear <strong>and</strong> disgust. In order to determine the categories of interest most of the studies focus<strong>on</strong> basic emoti<strong>on</strong>s since they are the most relevant because of their social or biological value butalso because they can be used to determine more complex n<strong>on</strong>-basic emoti<strong>on</strong>s. However, severallists of basic emoti<strong>on</strong>s are available (Table 2.1 is just a sample of those lists) <strong>and</strong> it is thusnecessary to make choices. For instance, Ekman’s six basic emoti<strong>on</strong>s, also known as the big six,have been widely used as the target classes to be recognized from the analysis of facialexpressi<strong>on</strong>s [20, 65-68]. Those target classes were chosen because Ekman’s definiti<strong>on</strong> of basicemoti<strong>on</strong>s was based <strong>on</strong> the universality of facial expressi<strong>on</strong>s. Having a system able to detect thebig six from facial expressi<strong>on</strong>s would thus be universal. But what happens if <strong>on</strong>e wants to assessemoti<strong>on</strong>s based <strong>on</strong> others modalities such as signals of the body? There is evidence that some ofEkman’s basic emoti<strong>on</strong>s have specific patterns of physiological activity [5] but it is still not clearif this is true <str<strong>on</strong>g>for</str<strong>on</strong>g> all of them. If brain activity is m<strong>on</strong>itored <strong>on</strong>e could rely <strong>on</strong> James’s basicemoti<strong>on</strong>s or <strong>on</strong> Gray <strong>and</strong> Panksepp sets (Table 2.1). But does it really make sense to change theset of targeted emoti<strong>on</strong>s according to the m<strong>on</strong>itored activity? A soluti<strong>on</strong> could be to have a lookat all lists of basic emoti<strong>on</strong>s <strong>and</strong> retrieve the most frequent <strong>on</strong>es to c<strong>on</strong>struct our own set ofemoti<strong>on</strong> to recognize. Un<str<strong>on</strong>g>for</str<strong>on</strong>g>tunately this is a difficult task because of the high number of suchlists <strong>and</strong> the ambiguity of emoti<strong>on</strong>al words (ie. are rage <strong>and</strong> anger different emoti<strong>on</strong>s or do theyrefer to the same state?).One interesting point coming from the psychologist view of basic emoti<strong>on</strong>s is the idea of mixtureof emoti<strong>on</strong>s. This idea states that emoti<strong>on</strong>s do not always appear in isolati<strong>on</strong> but that they canalso co-occur <strong>on</strong> a relatively short time scale (order of the sec<strong>on</strong>d). The co-occurred emoti<strong>on</strong>scould be similar but also opposite, <strong>and</strong> imply the activati<strong>on</strong> of two parallel, or at least two quicklyc<strong>on</strong>secutive appraisal processes. For instance, I used to visit my gr<strong>and</strong>-parents at the hospital <strong>and</strong>each time I entered the room, the elicited feeling was a mixture of pleasure <strong>and</strong> sadness. Pleasurebecause I was happy to see them again, sadness because I could see that they were not as healthyas I left them from the previous visit. There are two possible ways to account <str<strong>on</strong>g>for</str<strong>on</strong>g> such aphenomen<strong>on</strong> in emoti<strong>on</strong> assessment: the first is to c<strong>on</strong>sider that the mixture is a new emoti<strong>on</strong> (<str<strong>on</strong>g>for</str<strong>on</strong>g>instance by using Plutchik mixture model) <strong>and</strong> the sec<strong>on</strong>d is simply to allow <str<strong>on</strong>g>for</str<strong>on</strong>g> multiple labelingof emoti<strong>on</strong> episodes. In both cases it is necessary to first identify several emoti<strong>on</strong>s from <strong>on</strong>eepoch’s data. Several models have been proposed to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m multiple emoti<strong>on</strong> labeling [69, 70];of interest is the fact that some of those also allow <str<strong>on</strong>g>for</str<strong>on</strong>g> the characterizati<strong>on</strong> of intermediary states,24


State of the artoccurring at the transiti<strong>on</strong> of an emoti<strong>on</strong> to another, which can be regarded as mixed emoti<strong>on</strong>alstates [71].In any case, the definiti<strong>on</strong> of a set of emoti<strong>on</strong>al labels should be str<strong>on</strong>gly related to the targetedapplicati<strong>on</strong>. For instance, if the goal of an applicati<strong>on</strong> is to m<strong>on</strong>itor critical states while driving,there is actually no sense in detecting an emoti<strong>on</strong>al state such as disgust just because emoti<strong>on</strong>sare detected from facial expressi<strong>on</strong>. First, because there are very few chances that this emoti<strong>on</strong>arises during driving. Sec<strong>on</strong>dly, because there is no clear acti<strong>on</strong> that the system (the “emoti<strong>on</strong>allyintelligent” car) should take to deal with the situati<strong>on</strong>. On the other h<strong>and</strong> detecting boredom,which is not part of Ekman’s basic emoti<strong>on</strong>s, could be of interest.It is also important to check if the applicati<strong>on</strong> can gather c<strong>on</strong>textual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> that could beuseful <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment. For instance, the OCC model could be used together withc<strong>on</strong>textual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> to determine a path in the OCC tree while a binary assessment of valence(positive vs. negative) would then be enough to finally determine the elicited emoti<strong>on</strong>.More than detecting emoti<strong>on</strong> categories this approach could be useful to obtain in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>c<strong>on</strong>cerning the intensity of the felt emoti<strong>on</strong> which discrete labels generally fail to do. The labelsproposed by Plutchik in his wheel of emoti<strong>on</strong>s are related to intensity but this representati<strong>on</strong> hasthe disadvantage of increasing the number of emoti<strong>on</strong>s to detect (Figure 2.1). Another possibilityis to associate to each emoti<strong>on</strong>al label a c<strong>on</strong>tinuous level that will represent the intensity of eachemoti<strong>on</strong>. This is what Rani [70, 72] proposed after observing that emoti<strong>on</strong>al physiologicalreacti<strong>on</strong>s where different <str<strong>on</strong>g>for</str<strong>on</strong>g> low <strong>and</strong> high anger. Rani’s representati<strong>on</strong> has the advantage thatmixtures of emoti<strong>on</strong>s can also be represented. However, in this representati<strong>on</strong> it is important toaccount <str<strong>on</strong>g>for</str<strong>on</strong>g> the relati<strong>on</strong>s between the different emoti<strong>on</strong>s in order to avoid inc<strong>on</strong>sistent state suchas high anger <strong>and</strong> high boredom which is not the case in Rani’s proposal.As stated be<str<strong>on</strong>g>for</str<strong>on</strong>g>e, the c<strong>on</strong>tinuous representati<strong>on</strong> directly h<strong>and</strong>les the noti<strong>on</strong> of intensity. But itsuffers <str<strong>on</strong>g>for</str<strong>on</strong>g> a major drawback: it is not intuitive compared to discrete emoti<strong>on</strong>al labels leading todifficult manual tagging of emoti<strong>on</strong>al data. While this is not a very important issue if this task isd<strong>on</strong>e by an expert, most of tagging is actually d<strong>on</strong>e by people who do not have any knowledgeabout this type of representati<strong>on</strong>s. To alleviate this problem, a descripti<strong>on</strong> of how to use of avalence-arousal grid <str<strong>on</strong>g>for</str<strong>on</strong>g> self-assessment has been proposed in [52]. Alternatively, the SAM (Self<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> Manikin) has been proposed in [73]. The SAM is composed of three nine-pointscales where the different values of valence, arousal <strong>and</strong> dominance are associated withpictograms (Figure 2.5). Even if pictograms help the user to better underst<strong>and</strong> the definiti<strong>on</strong> ofthe different axes they should be introduced to him / her prior to self-assessment. Anotherdisadvantage of the valence-arousal space is that mixtures of negative <strong>and</strong> positive emoti<strong>on</strong>scannot be represented. It is then necessary to decompose the emoti<strong>on</strong> into two points in the25


Chapter 2c<strong>on</strong>tinuous space or choose the average resp<strong>on</strong>se. This often c<strong>on</strong>fuses users when they try to selfassesstheir emoti<strong>on</strong>s <strong>on</strong> a l<strong>on</strong>g time scale where there are probably several emoti<strong>on</strong>s that can beelicited.One of the main advantages of c<strong>on</strong>tinuous spaces is that it can be seen as a general representati<strong>on</strong>of any emoti<strong>on</strong>. As a result it is possible to use them in any applicati<strong>on</strong>, with <strong>on</strong>ly the area ofinterest varying from an applicati<strong>on</strong> to another. In the case of the valence-arousal space it seemsthat <strong>on</strong>ly some areas are of real importance [23, 54]: negative-excited, positive-excited <strong>and</strong> calmneutral (Figure 2.3). Another advantage is that if a point in this space is determined (by automaticemoti<strong>on</strong> assessment <str<strong>on</strong>g>for</str<strong>on</strong>g> instance), it is possible to associate an emoti<strong>on</strong>al label to it. It is alsopossible to assess <strong>on</strong>ly <strong>on</strong>e axis of the valence-arousal space which could be enough <str<strong>on</strong>g>for</str<strong>on</strong>g> certainapplicati<strong>on</strong>s (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance, m<strong>on</strong>itoring of arousal <str<strong>on</strong>g>for</str<strong>on</strong>g> driving). Because of its generality but also<str<strong>on</strong>g>for</str<strong>on</strong>g> the other reas<strong>on</strong>s menti<strong>on</strong>ed above, we believe that it is preferable to assess valence-arousalspace dimensi<strong>on</strong>s or areas than discrete labels if there is no particular targeted applicati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g>emoti<strong>on</strong> assessment.Figure 2.5. Picture of the SAM scales (from [73]).The first line evaluates valence from positive (left) t<strong>on</strong>egative (right), the sec<strong>on</strong>d arousal from excited to calm <strong>and</strong> the third dominance from submissive topowerful.2.2 Physiological signalsAs sustained by the SECs theory, emoti<strong>on</strong>s should be assessed from the m<strong>on</strong>itoring of the fivesystems involved in emoti<strong>on</strong> elicitati<strong>on</strong> (see Secti<strong>on</strong> 2.1.3.b). In this thesis, the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mances oftwo of those systems are analyzed <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment: the cognitive <strong>and</strong> the aut<strong>on</strong>omicsystems. Several signals can be gathered to relate activities of the five systems (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance acamera can be used to partially measure the motor activity) but the present work focuses <strong>on</strong>signals recorded by n<strong>on</strong>-invasive sensors placed directly <strong>on</strong> the body: the physiological signals.26


State of the artPhysiological signals can be defined as signals that quantify physical <strong>and</strong> chemical phenomenaoccurring in organs <strong>and</strong> tissues. This definiti<strong>on</strong> emphasizes the existence of a device able toquantify a physical <strong>and</strong> chemical phenomen<strong>on</strong>: the sensor. The complete apparatus allowing thevisualizati<strong>on</strong> <strong>and</strong> recording of (many) physiological signals is called an acquisiti<strong>on</strong> system <strong>and</strong> isdetailed in Figure 2.6.Physical <strong>and</strong>chemicalphenomenaSensorsAnalog electricalsignalsAnalog. / Digitalc<strong>on</strong>verterStorageTemperaturePressureElectrical potentials...Digital signalsVisualizati<strong>on</strong>Figure 2.6. An acquisiti<strong>on</strong> system <str<strong>on</strong>g>for</str<strong>on</strong>g> visualizati<strong>on</strong> <strong>and</strong> storage of physiological data.There is a wide variety of physiological signals depending <strong>on</strong> their source (the organ underc<strong>on</strong>siderati<strong>on</strong>) <strong>and</strong> the phenomen<strong>on</strong> that is measured. For instance heart activity can be measuredby recording the electrical potentials <strong>on</strong> particular positi<strong>on</strong>s of the body, the heart sound <strong>and</strong> theblood pressure in the cor<strong>on</strong>ary arteries. In this thesis, since all organs are c<strong>on</strong>nected to <strong>and</strong>c<strong>on</strong>trolled by the nervous system, we chose to divide them according to the following tax<strong>on</strong>omy:the central nervous system (CNS) <strong>and</strong> the peripheral nervous system (PNS). Thus, activities o<str<strong>on</strong>g>for</str<strong>on</strong>g>gans that are c<strong>on</strong>trolled by the PNS will be referred to as peripheral activities while the activityof other organs (in this study <strong>on</strong>ly the brain) will be named the central activity. Notice that thistax<strong>on</strong>omy enables us to separate the two systems that are supposed to be at the center of thecognitive <strong>and</strong> Jamesian theories of emoti<strong>on</strong>s.2.2.1 Central nervous system (CNS)The CNS is composed of the brain, the cerebellum, the brain stem <strong>and</strong> the spinal cord. Since inthis study <strong>on</strong>ly brain activity was analyzed, this descripti<strong>on</strong> will focus <strong>on</strong> the brain structures <strong>and</strong>activities related to emoti<strong>on</strong>s as well as <strong>on</strong> the devices that could be used to measure thoseactivities.a. <strong>Brain</strong> structureTo better underst<strong>and</strong> the functi<strong>on</strong>ing of the different devices used to m<strong>on</strong>itor brain activity <strong>and</strong>the analysis of the resulting signals, it is first necessary to briefly discuss the anatomy of thebrain. The human brain is made of around 100 billi<strong>on</strong>s of neur<strong>on</strong>s that are interc<strong>on</strong>nected throughsynapses, c<strong>on</strong>stituting neur<strong>on</strong>al networks. Neur<strong>on</strong>s are specialized cells that vary in size <strong>and</strong>27


Chapter 2shapes but are always c<strong>on</strong>stituted of three main parts: the dendrites, the soma <strong>and</strong> the ax<strong>on</strong>(Figure 2.7.a). The soma is the cell body of the neur<strong>on</strong> <strong>and</strong> c<strong>on</strong>tains the same elements as typicalcells (nucleus, ribosomes, etc.). The dendrites receive the output of <strong>on</strong>e or many incidentalneur<strong>on</strong>s as inputs while the ax<strong>on</strong> propagate the output of its neur<strong>on</strong>. The electrical signaltransmitted from <strong>on</strong>e neur<strong>on</strong> to another is composed of acti<strong>on</strong> potentials. The main task of aneur<strong>on</strong> is to integrate input acti<strong>on</strong> potentials to determine if an acti<strong>on</strong> potential should begenerated <strong>on</strong> the ax<strong>on</strong> (Figure 2.7.b). The activity of a neur<strong>on</strong> is proporti<strong>on</strong>al to its firing rate: themore a neur<strong>on</strong> is active the higher the firing rate.Ax<strong>on</strong> ofneur<strong>on</strong> 1Ax<strong>on</strong> ofneur<strong>on</strong> 2SynapseDendritesSomaAx<strong>on</strong>eNeur<strong>on</strong> 1Input acti<strong>on</strong>potentialsNeur<strong>on</strong> 2timetimeActi<strong>on</strong>potentialemissi<strong>on</strong>a) b)timeFigure 2.7. (a) Figure of a neur<strong>on</strong> 3 c<strong>on</strong>nected with two input neur<strong>on</strong>s (named 1 <strong>and</strong> 2). (b) Representati<strong>on</strong> ofthe integrati<strong>on</strong> of input acti<strong>on</strong> potentials; the neur<strong>on</strong> fires <strong>on</strong>ly if its membrane potential exceeds a giventhreshold.The brain is divided into four lobes according to the names of the b<strong>on</strong>es that surround them <strong>and</strong>the sulcus that separate the different lobes (Figure 2.8): the fr<strong>on</strong>tal, parietal, temporal <strong>and</strong>occipital lobes. Each of those lobes has been c<strong>on</strong>sidered to be specialized in particular cognitivetasks. For instance, the fr<strong>on</strong>tal lobe is known to be implied in planning tasks while the occipitallobe is c<strong>on</strong>sidered to be the visi<strong>on</strong> center. Even if this view is superseded by new approaches thatemphasize the cooperati<strong>on</strong> between different areas of the brain during a single cognitive task, thisnomenclature is still used to describe areas of the brain.3 http://www.wiredtowinthemovie.com/mindtrip_xml.html (retrieved <strong>on</strong> 27 April 2009)28


State of the artFr<strong>on</strong>tal lobeParietal lobeOccipital lobeTemporal lobe<strong>Brain</strong> stemCerebellumFigure 2.8. Image of the brain, the brain stem <strong>and</strong> the cerebellum with the different lobes highlighted (from[74]).b. Measuring brain activitySince neur<strong>on</strong>s communicate trough acti<strong>on</strong> potentials it is possible to m<strong>on</strong>itor brain activity bymeasuring electro-magnetic fields. Those methods are known as direct methods since theydirectly measure the activity produced by the neur<strong>on</strong>s <strong>and</strong> can be divided into two types:- electroencephalography (EEG) measures electrical potentials by placing electrodes either<strong>on</strong> the scalp (surface EEG) or, by incising the skull, <strong>on</strong> the surface of the brain (Electro-Cortico-Graphic electrodes - ECoG) or directly in neur<strong>on</strong>s of interests (intra-corticalelectrodes);- magnetoencephalography (MEG) measures the magnetic activity of neur<strong>on</strong>s by usingspecialized devices (Superc<strong>on</strong>ducting QUantum Interference Devices - SQUID) able todetect small changes in magnetic fields.Indirect methods m<strong>on</strong>itor other parameters that are related to neur<strong>on</strong>al activity such as artificialtracers injected in the body <strong>and</strong> resources c<strong>on</strong>sumed by neur<strong>on</strong>s (oxygen <str<strong>on</strong>g>for</str<strong>on</strong>g> instance):- Positr<strong>on</strong> Emissi<strong>on</strong> Tomography (PET) measures the positr<strong>on</strong> emissi<strong>on</strong> of a slightlyradioactive tracer;- Single Phot<strong>on</strong> Emissi<strong>on</strong> Computed Tomography (SPECT), similarly to the PET, requiresthe injecti<strong>on</strong> of a radioactive tracer emitting gamma radiati<strong>on</strong>s that are measured by agamma ray camera;- functi<strong>on</strong>al Magnetic Res<strong>on</strong>ance Imagery (fMRI) detects the changes of oxyhemoglobin<strong>and</strong> desoxyhemoglobin by applying successive magnetic fields <strong>on</strong> the head that <str<strong>on</strong>g>for</str<strong>on</strong>g>ceprot<strong>on</strong>s to release energy at a particular frequency that is detectable by the system;29


Chapter 2- functi<strong>on</strong>al Near InfraRed Spectroscopy (fNIRS) also m<strong>on</strong>itors the changes of oxydesoxyhemoglobinby measuring the reflectance of a near infrared diode (0.7-1.5 µmwave-length) placed <strong>on</strong> the scalp using a near infrared sensor placed next to the diode.All those methods can be evaluated according to several criteria: spatial resoluti<strong>on</strong>, timeresoluti<strong>on</strong>, financial cost, needed machinery size / weight, invasiveness, chirurgical operati<strong>on</strong><strong>and</strong>/or technical preparati<strong>on</strong>. Table 2.3 evaluates the different methods above according to thosecriteria. High spatial <strong>and</strong> temporal resoluti<strong>on</strong>s are important <str<strong>on</strong>g>for</str<strong>on</strong>g> a precise analysis of signals butcurrent methods do not allow <str<strong>on</strong>g>for</str<strong>on</strong>g> both high spatial <strong>and</strong> temporal resoluti<strong>on</strong>. However, recentstudies are trying to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m multimodal recording of brain activity in order to combineresoluti<strong>on</strong> advantages of direct <strong>and</strong> indirect methods.Direct methodsMethodsSurfaceEEGSpatialresoluti<strong>on</strong>Timeresoluti<strong>on</strong>FinancialcostMachinerysize / weightInvasive.Low High Low Low N<strong>on</strong>eECoG Low High High Low HighIntracorticalHigh High High Low HighMEG Low High High High N<strong>on</strong>ePreparati<strong>on</strong>Apply gel <strong>on</strong> thescalpSurgical incisi<strong>on</strong>of the skullSurgical incisi<strong>on</strong>of the skull <strong>and</strong>inserti<strong>on</strong> of theelectrode in thebrainIndirectmethodsPET High Low High High LowInjecti<strong>on</strong> of aradioactive tracerSPECT Low Low High High LowInjecti<strong>on</strong> of aradioactive tracerfMRI High Low High High N<strong>on</strong>efNIRS Low Low Low Low N<strong>on</strong>eTable 2.3. Comparis<strong>on</strong> of the different methods <str<strong>on</strong>g>for</str<strong>on</strong>g> the m<strong>on</strong>itoring of brain activity.The criteria can also be viewed as c<strong>on</strong>straints <str<strong>on</strong>g>for</str<strong>on</strong>g> the use of devices in HMI. For instance, arelatively low financial cost is m<strong>and</strong>atory <str<strong>on</strong>g>for</str<strong>on</strong>g> a commercial use of an emoti<strong>on</strong> detecti<strong>on</strong> system.Having devices that are easily wearable <strong>and</strong> n<strong>on</strong> invasive is also essential since it is unlikely thatusers will accept to receive injecti<strong>on</strong>s or have a heavy chirurgical operati<strong>on</strong> just <str<strong>on</strong>g>for</str<strong>on</strong>g> using a newHMI. For this reas<strong>on</strong>, as can be seen from Table 2.3, surface EEG is certainly the most applicablemethod <str<strong>on</strong>g>for</str<strong>on</strong>g> use in HMI systems. As a c<strong>on</strong>sequence, this sensor was chosen <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> detecti<strong>on</strong>in this study. However EEG still requires quite some preparati<strong>on</strong>, such as the applicati<strong>on</strong> of ac<strong>on</strong>ductive gel <strong>on</strong> the surface of the scalp to insure c<strong>on</strong>tact between the skin <strong>and</strong> electrodes, <strong>and</strong>the plugging of each electrode in a tightly fitting headcap. This can be seen as a high c<strong>on</strong>straint togo toward commercial applicati<strong>on</strong>s but there are now wireless dry-electrodes systems available30


State of the art<strong>on</strong> the market which do not require as much preparati<strong>on</strong> <strong>and</strong> are quite cheap (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance theNeuroSky 4 <strong>and</strong> Emotiv 5 devices).Since in this thesis surface EEGs were used to m<strong>on</strong>itor brain activity, the term EEG will now <strong>on</strong>refer to this type of measurement <strong>and</strong> not to EcoG <strong>and</strong> intra-cortical measurements. The electricalactivity of a neur<strong>on</strong> has first to go trough several layers of different matter to reach an EEGelectrode (gray matter, white matter, cerebrospinal fluid, b<strong>on</strong>e <strong>and</strong> skin). The signal measured byan EEG electrode is thus the integrati<strong>on</strong> of the signals emitted by all the neur<strong>on</strong>s that are closeenough to be recorded. However, researchers have found that there are particular frequency b<strong>and</strong>s(often called rhythms in the literature) of interest to interpret EEG signals: delta (2-4 Hz), theta(4-8 Hz), alpha (8-12 Hz), beta (12-30 Hz) <strong>and</strong> gamma (> 30 Hz) b<strong>and</strong>s. Notice that thefrequency b<strong>and</strong>s associated to each rhythm can vary from a study to another <strong>and</strong> from <strong>on</strong>e pers<strong>on</strong>to another. Moreover, different names can be given to frequency b<strong>and</strong>s according to the functi<strong>on</strong>that is associated to it, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance the alpha b<strong>and</strong> is also referred to as the mu-rhythm whenc<strong>on</strong>sidering the activity in the motor cortex. The reas<strong>on</strong> why those rhythms are observed <strong>and</strong>fluctuate is still unclear but it is supposed that synchr<strong>on</strong>izati<strong>on</strong> of several populati<strong>on</strong>s of neur<strong>on</strong>scould be at the source of those phenomena [75].Currently, EEG systems can record signals with more than 250 electrodes. Increasing the numberof electrodes can improve the spatial resoluti<strong>on</strong> but increases the time necessary to apply gel <strong>and</strong>plug electrodes as well as the number of variables to analyze. In order <str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG community tohave some st<strong>and</strong>ards c<strong>on</strong>cerning the positi<strong>on</strong>ing <strong>and</strong> the naming of electrodes the 10-20 systemwas proposed [76]. Its name comes from the fact that the fr<strong>on</strong>t <strong>and</strong> back median electrodes arepositi<strong>on</strong>ed at 10% of the ini<strong>on</strong>-nasi<strong>on</strong> distance while the other electrodes are separated from eachother by 20% of the same distance. The ini<strong>on</strong> is a b<strong>on</strong>e <strong>on</strong> the back of the skull <strong>and</strong> the nasi<strong>on</strong> isthe intersecti<strong>on</strong> of b<strong>on</strong>es just above the nose. Latter, other systems extended the 10-20 system todefine locati<strong>on</strong>s <strong>and</strong> names of more electrodes: the 10-10 system (74 locati<strong>on</strong>s) <strong>and</strong> the 10-5system (345 locati<strong>on</strong>s) [76].Noise is a critical issue in the measurement of EEG. According to Kr<strong>on</strong>egg [28], there are at leastthree potential sources of noises. Envir<strong>on</strong>ment noises c<strong>on</strong>cern all the electrical noises thatsurround the recording, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance the 50Hz power line noise. Physiological noises refer t<strong>on</strong>oises from the body such as muscles c<strong>on</strong>tracti<strong>on</strong>s. The last source of noise is the backgroundactivity of the brain <strong>and</strong> is defined as all brain activities that are recorded but are not related to theparticular brain functi<strong>on</strong> under study.4 http://www.neurosky.com/ (retrieved <strong>on</strong> 29 April 2009)5 http://emotiv.com/ (retrieved <strong>on</strong> 29 April 2009)31


Chapter 2c. The brain <strong>and</strong> emoti<strong>on</strong>sThe cognitive view of emoti<strong>on</strong>s supports the idea that brain structures are involved in emoti<strong>on</strong>alprocesses. To gather evidences supporting this hypothesis, researchers have first analyzed thebehavior of patients suffering from brain damages <strong>and</strong> animals with segmented brain areas. Inthis way they could identify structures that regulate emoti<strong>on</strong>al expressi<strong>on</strong>s. The inventi<strong>on</strong> of newmethods to m<strong>on</strong>itor brain activity such as EEG <strong>and</strong> fMRI allowed <str<strong>on</strong>g>for</str<strong>on</strong>g> a more precise <strong>and</strong> easieridentificati<strong>on</strong> of those structures as well as <str<strong>on</strong>g>for</str<strong>on</strong>g> the determinati<strong>on</strong> of their functi<strong>on</strong>s in emoti<strong>on</strong>alprocesses.One of the first structures that was shown to be str<strong>on</strong>gly involved in emoti<strong>on</strong>al expressi<strong>on</strong> is thelimbic system. The elements composing this system are part of the temporal lobes <strong>and</strong> are placedunder the cortex, deep in the brain. Papez was the first to suspect this system to be partiallydedicated to emoti<strong>on</strong>s <strong>and</strong> found that the elements of the limbic system <str<strong>on</strong>g>for</str<strong>on</strong>g>m a circuit, called thePapez circuit, that interc<strong>on</strong>nect the cortex with the hypothalamus through structures such as thedorsal thalamus, the cingular gyrus, the hippocampus <strong>and</strong> the <str<strong>on</strong>g>for</str<strong>on</strong>g>nix [74]. Since the hypothalamuswas known to coordinate emoti<strong>on</strong>al behaviors <strong>and</strong> expressi<strong>on</strong>s [74] this discovery supports theexistence of a higher level of emoti<strong>on</strong>al processing occurring in the cortex. Later, an importantstructure was added as a part of this circuit: the amygdala [77]. It has been found that theamygdala is highly interc<strong>on</strong>nected with sensory cortical areas suggesting that it could be at thesource of the emoti<strong>on</strong>al attributi<strong>on</strong>s to sensorial stimuli (Figure 2.9). This hypothesis has beensupported by many studies that shows the activati<strong>on</strong> of this regi<strong>on</strong> at least <str<strong>on</strong>g>for</str<strong>on</strong>g> negative emoti<strong>on</strong>ssuch as fear <strong>and</strong> anger [78]. It was also proposed that the amygdala is involved in rein<str<strong>on</strong>g>for</str<strong>on</strong>g>cementlearning [78, 79] <strong>and</strong> recogniti<strong>on</strong> of emoti<strong>on</strong>s [80].StimulusLimbic systemHypotalamusCoordinate emoti<strong>on</strong>albehaviorPeripheralnervous systemExecute behaviorCortexProcess sensorialin<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>sAnterior <strong>and</strong> cingularcortexAssociative f<strong>on</strong>cti<strong>on</strong>sAmygdalaProcess emoti<strong>on</strong>alsignificance (<strong>and</strong> per<str<strong>on</strong>g>for</str<strong>on</strong>g>mrein<str<strong>on</strong>g>for</str<strong>on</strong>g>cement learning)Figure 2.9. Principal structures of the limbic system together with their functi<strong>on</strong>s.The other brain area that is assumed to play a key role in the elicitati<strong>on</strong> of emoti<strong>on</strong>s is theprefr<strong>on</strong>tal cortex [74, 78-81]. This area is located at the fr<strong>on</strong>t of the fr<strong>on</strong>tal lobe <strong>and</strong> c<strong>on</strong>tains theorbito-fr<strong>on</strong>tal cortex (above the orbits of the eyes) that is known to be involved in cognitive32


State of the artprocesses such as decisi<strong>on</strong> making. Davids<strong>on</strong> [19, 81] has shown that the prefr<strong>on</strong>tal cortex isinvolved in approach <strong>and</strong> withdrawal reacti<strong>on</strong>s, which are closely linked with emoti<strong>on</strong>s, whileRolls [79] argues that the orbito-fr<strong>on</strong>tal cortex also plays a role in rein<str<strong>on</strong>g>for</str<strong>on</strong>g>cement learning ofemoti<strong>on</strong>s.Lateralizati<strong>on</strong> of those brain structures is also of interest. The most c<strong>on</strong>sistent findingsc<strong>on</strong>cerning this hypothesis c<strong>on</strong>cern the prefr<strong>on</strong>tal cortex. At least two types of lateralizati<strong>on</strong> havebeen observed. First Davids<strong>on</strong> [19, 81] has noticed that the left prefr<strong>on</strong>tal lobe is more involvedin emoti<strong>on</strong>al reacti<strong>on</strong>s corresp<strong>on</strong>ding to approach reacti<strong>on</strong>s while the right prefr<strong>on</strong>tal lobe is moreinvolved in withdrawal reacti<strong>on</strong>s. This was c<strong>on</strong>cluded by observing the power changes in thealpha b<strong>and</strong> from EEG signals of participants subject to the two different types of stimuli. Someresearchers also suggest that this lateralizati<strong>on</strong> corresp<strong>on</strong>ds to positive <strong>and</strong> negative emoti<strong>on</strong>s.However, this theory is discussed since the relati<strong>on</strong> between approach-withdrawal reacti<strong>on</strong>s <strong>and</strong>positive-negative emoti<strong>on</strong>s is not direct. For instance, some negative stimuli can lead to approachreacti<strong>on</strong>s (anger <str<strong>on</strong>g>for</str<strong>on</strong>g> instance). Sec<strong>on</strong>dly, evidences suggest that the right hemisphere is moreinvolved in emoti<strong>on</strong>al processing than the left <strong>on</strong>e. Some studies also argue <str<strong>on</strong>g>for</str<strong>on</strong>g> asymmetricphenomen<strong>on</strong> in the amygdala but this questi<strong>on</strong> is still under study [78].Finally, there is a high number of studies showing the activati<strong>on</strong> of several other brain areasduring emoti<strong>on</strong>al processes. An example is the work of Aftanas et al. [82] that showed significantdifferentiati<strong>on</strong> of arousal based <strong>on</strong> EEG data collected from participants watching high,intermediate <strong>and</strong> low arousal images. This differentiati<strong>on</strong> could be found in various areas such asparietal, parieto-temporal <strong>and</strong> occipital lobes. In [83] interacti<strong>on</strong>s between cortical regi<strong>on</strong>s duringthe presentati<strong>on</strong> of emoti<strong>on</strong>al film clips were analyzed. It turned out that brain areas weredifferently synchr<strong>on</strong>ized depending <strong>on</strong> the type of stimulus. For instance, a highersynchr<strong>on</strong>izati<strong>on</strong> between left <strong>and</strong> right fr<strong>on</strong>tal areas was observed <str<strong>on</strong>g>for</str<strong>on</strong>g> sad clips when compared tohappy clips. Damasio also observed differences in brain activity during the feeling of selfgeneratedemoti<strong>on</strong>s [84].2.2.2 Peripheral nervous system (PNS)The PNS is c<strong>on</strong>stituted of neur<strong>on</strong>s that can be of two types: the sensory neur<strong>on</strong>s that c<strong>on</strong>veyin<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> from the sensory receptors to the CNS <strong>and</strong> the motor neur<strong>on</strong>s that arouse or inhibitmuscles <strong>and</strong> gl<strong>and</strong>s activity. The PNS can be divided in two subsystems: the somatic nervoussystem (SNS) that is c<strong>on</strong>nected to skeletal muscles <strong>and</strong> the aut<strong>on</strong>omic nervous system (ANS)c<strong>on</strong>nected to aut<strong>on</strong>omous muscles (the heart <str<strong>on</strong>g>for</str<strong>on</strong>g> instance) <strong>and</strong> gl<strong>and</strong>s. The SNS is associated withvoluntary c<strong>on</strong>trol of muscles while the ANS does not require c<strong>on</strong>scious c<strong>on</strong>trol to activate orinhibit c<strong>on</strong>nected organs. The SNS <strong>and</strong> the ANS are both involved in emoti<strong>on</strong>al reacti<strong>on</strong>s. Forinstance, the SNS is c<strong>on</strong>cerned with the voluntary producti<strong>on</strong> of facial expressi<strong>on</strong>s <strong>and</strong> the ANS33


Chapter 2with involuntary facial expressi<strong>on</strong>s <strong>and</strong> physiological supportive functi<strong>on</strong>s (see secti<strong>on</strong> 2.1.3.b).Finally the ANS is also divided into two categories: the sympathetic nervous system <strong>and</strong> theparasympathetic nervous system. While the first prepares the organism <str<strong>on</strong>g>for</str<strong>on</strong>g> the uses of resourcesgenerally related to “flight <strong>and</strong> fight” resp<strong>on</strong>ses, the sec<strong>on</strong>d has the role of preserving <strong>and</strong>augmenting resources. Both of those systems are thus of interest in the study of emoti<strong>on</strong>s.Since it is difficult <strong>and</strong> invasive to measure PNS electrical potentials by inserting electrodesdirectly into the nerves, PNS activity is generally indirectly assessed by measuring the activity ofperipheral organs such as muscles. Below are described different measures of the PNS that willbe used in this study.a. Electrodermal activityThe ElectroDermal Activity (EDA) is related to the changes of electrical potentials <strong>and</strong> resistanceof the skin. It was first measured by Féré [85]. While phenomen<strong>on</strong> inducing electrodermalresp<strong>on</strong>ses are still not completely understood, it has been shown that they are str<strong>on</strong>gly associatedwith sweat gl<strong>and</strong> activity [86]. Sweat gl<strong>and</strong>s are situated deeply in the skin at the juncti<strong>on</strong> of thehypodermis <strong>and</strong> the dermis. A duct links a sweat gl<strong>and</strong> to a pore situated at the surface of the skin(the epidermis) that can be closed or opened to let sweat spread out. Since sweat gl<strong>and</strong> activity isknown to be c<strong>on</strong>trolled by the sympathetic nervous system, EDA has become a comm<strong>on</strong> sourceof in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> to measure the ANS.There are two methods to measure EDA:- the exosomatic method c<strong>on</strong>sist of placing active electrodes (electrodes that inject a smallcurrent) <strong>on</strong> two sites of the skin to measure its resistance (or its c<strong>on</strong>ductance);- the endosomatic method c<strong>on</strong>sists of placing two electrodes <strong>on</strong> the skin to simply measurethe differences in skin potentials.In both cases, since the sweat is a salty liquid that plays the role of a c<strong>on</strong>ductor, as so<strong>on</strong> as theducts are filled with it, the electrical <strong>and</strong> impedance properties of the skin will change <strong>and</strong> berecorded by <strong>on</strong>e the above settings. The palms of h<strong>and</strong>s <strong>and</strong> foot, especially the fingers <strong>and</strong> thetoes, are recommended places to positi<strong>on</strong> the electrodes because at these areas the skin c<strong>on</strong>tainsmany sweat gl<strong>and</strong>s.The measured EDA can be decomposed in two different comp<strong>on</strong>ents [87]: the t<strong>on</strong>ic level <strong>and</strong> thephasic resp<strong>on</strong>se as highlighted in Figure 2.10. The t<strong>on</strong>ic level represents the general resistance ofthe skin <strong>and</strong> is influenced by the hydrati<strong>on</strong> of the skin <strong>and</strong> the precedent sweat emissi<strong>on</strong>s. Thephasic resp<strong>on</strong>se is due to the accumulati<strong>on</strong> of sweat in ducts <strong>and</strong> opening of the pores. Notice thatit is not necessary that the sweat reaches the surface of the skin <str<strong>on</strong>g>for</str<strong>on</strong>g> the resp<strong>on</strong>se to occur. This34


State of the artresp<strong>on</strong>se occurs from 1 to 4 sec<strong>on</strong>ds after a stimulus [88] <strong>and</strong> is often characterized by its latency(the time <str<strong>on</strong>g>for</str<strong>on</strong>g> the resp<strong>on</strong>se to start after the stimuli), amplitude, rise time <strong>and</strong> half-recovery time(Figure 2.10) [87]. Sometimes these characteristics are hard to determine when there are severalresp<strong>on</strong>ses that overlap.Depending <strong>on</strong> the method used to measure EDA as well as <strong>on</strong> the comp<strong>on</strong>ent of interest, severalnames can be found in the literature that refer to the same measurements: ElectroDermalResp<strong>on</strong>se (EDR), Galvanic Skin Resp<strong>on</strong>se (GSR), Skin Resistance Level / Resp<strong>on</strong>se (SRL <strong>and</strong>SRR), Skin C<strong>on</strong>ductance Level / Resp<strong>on</strong>se (SCL <strong>and</strong> SCR), Skin Potential Level / Resp<strong>on</strong>se(SPL <strong>and</strong> SPR).An increase of activity in sweat gl<strong>and</strong>s usually occurs when <strong>on</strong>e is experiencing emoti<strong>on</strong>s such asstress or surprise. By asking participants to look at more or less arousing pictures of the IAPS,Lang et al. [7] discovered that the SCR amplitude is correlated with the level of arousal. Similarresults where obtained in [89] showing that emoti<strong>on</strong>al words <strong>and</strong> neutral words can bedifferentiated based an EDA measures. By analyzing reacti<strong>on</strong>s to olfactory stimuli, Delplanque etal. [90] found a higher amplitude of the EDR <str<strong>on</strong>g>for</str<strong>on</strong>g> unpleasant than <str<strong>on</strong>g>for</str<strong>on</strong>g> pleasant odors. Moreover,EDA is known to be influenced by brain structures such as the hypothalamus, the limbic system<strong>and</strong> fr<strong>on</strong>tal cortical areas [89]. EDA can thus also be regarded as a window <strong>on</strong> emoti<strong>on</strong>al brainactivity.T<strong>on</strong>ic levelSkin resistance (Ohms)Overlap oftwo resp<strong>on</strong>sesElectrodermalresp<strong>on</strong>sesHalf recovery timeRise timeFigure 2.10. (left) Example of a signal representing the changes of resistance of the skin, (right) thecharacterizati<strong>on</strong> of an electrodermal resp<strong>on</strong>se.b. Blood pressureThe blood pressure (BP) is defined as the pressure that the blood exerts <strong>on</strong> the surfaces of thecardiovascular system (heart <strong>and</strong> vessels). Close to the heart the BP is the highest <strong>and</strong> decreasesas the blood flows in the cardiovascular system. Changes in BP are influenced by cardiac output35


Chapter 2(c<strong>on</strong>tracti<strong>on</strong>s of the heart), vasoc<strong>on</strong>stricti<strong>on</strong> <strong>and</strong> vasodilatati<strong>on</strong> (reducti<strong>on</strong> <strong>and</strong> augmentati<strong>on</strong> ofthe vessels diameter by c<strong>on</strong>tracti<strong>on</strong> <strong>and</strong> dilatati<strong>on</strong>), <strong>and</strong> also by mechanisms occurring in otherorgans [91].Generally two types of BP are distinguished: the systolic BP (SPB) is measured when the heart isc<strong>on</strong>tracted while the diastolic BP (DBP) is measured between two beats when the heart isrelaxed. In turn, the pulse pressure is defined as the difference between SBP <strong>and</strong> DBP. Anothervalue of interest is the mean arterial pressure (MAP) defined as the average BP. This value can becomputed by averaging c<strong>on</strong>tinuous measures of BP or by using an approximati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g>mula based<strong>on</strong> the discrete measures of SPB <strong>and</strong> DPB [91]. Finally, it is also possible to compute pulsetransit time (proporti<strong>on</strong>al to what is also called pulse wave velocity) defined as the time <str<strong>on</strong>g>for</str<strong>on</strong>g> thepressure wave to propagate from the heart to the area where the pressure is measured [91].In order to avoid invasive methods, BP is generally measured using indirect methods. They canbe divided in two types: intermittent methods that <strong>on</strong>ly allow <str<strong>on</strong>g>for</str<strong>on</strong>g> BP measures separated byseveral sec<strong>on</strong>ds <strong>and</strong> c<strong>on</strong>tinuous methods that more or less c<strong>on</strong>tinuously record the pressuresignal. Manual or automatic cuffs tied around the arm to measure brachial artery pressure(sphygmomanometers) are examples of intermittent methods that measure SBP <strong>and</strong> DBP byidentifying the so-called Korotkoff sounds. This method allows <str<strong>on</strong>g>for</str<strong>on</strong>g> 1 to 4 measures per minute[91]. By tying a photoplethysmograph around a finger it is possible to c<strong>on</strong>tinuously recordchanges in blood pressure. The photoplethysmograph is a device that emits light <strong>and</strong> measures itsreflecti<strong>on</strong> through the skin. Since light reflecti<strong>on</strong> varies according to the quantity of blood presentin the vessels this sensor is able to measure Blood Volume Pulse (BVP). The obtained signalcorrelates with BP but can <strong>on</strong>ly provide relative changes of BP. Usually this sensor is placed atareas where there are many vessels near to the surface of the skin such as fingers <strong>and</strong> ear lobes.Blood pressure has significant correlati<strong>on</strong> with defensive reacti<strong>on</strong>s since these reacti<strong>on</strong>s areassociated with vasoc<strong>on</strong>stricti<strong>on</strong> resp<strong>on</strong>ses [87]. Several emoti<strong>on</strong>s are also known to be related toan increase or decrease of cutaneous blood flow giving rise to flushes or paleness [74]. Sinha etal. [92] refers to the increase of blood pressure during fear <strong>and</strong> anger as <strong>on</strong>e of the mostc<strong>on</strong>sistent findings in emoti<strong>on</strong> research from aut<strong>on</strong>omic activity. Moreover, they reported anincrease in SBP during the visualizati<strong>on</strong> of emoti<strong>on</strong>al images compared to neutral images <strong>and</strong> anincrease in DBP <str<strong>on</strong>g>for</str<strong>on</strong>g> anger compared to sadness. Finally, Lisetti <strong>and</strong> Nasoz [93] listed in theirreview two studies [94, 95] where SBP <strong>and</strong> DBP were found to be reliable indicators of taskdifficulty.c. Heart rateThe Heart Rate (HR) is the number of heart c<strong>on</strong>tracti<strong>on</strong>s occurring over a given amount of time.It is generally expressed in Beats Per Minute (BPM) <strong>and</strong> is computed from InterBeat Intervals36


State of the art(IBI), the time elapsed between two beats. An increase of HR can be due to an increase of thesympathetic activity or a decrease of the parasympathetic activity. Moreover, the cardiacresp<strong>on</strong>se to sympathetic stimulati<strong>on</strong>s is slower than <str<strong>on</strong>g>for</str<strong>on</strong>g> parasympathetic stimulati<strong>on</strong>s giving riseto complex fluctuati<strong>on</strong>s of HR [96].To better analyze those fluctuati<strong>on</strong>s Heart Rate Variability (HRV) is a comm<strong>on</strong> feature extractedfrom HR <strong>and</strong> IBI. Several methods can be used to compute variables related to HRV. One of thesimplest is to c<strong>on</strong>sider the time series of IBI as a stochastic process <strong>and</strong> compute its statisticalproperties, such as st<strong>and</strong>ard deviati<strong>on</strong>, to characterize the distributi<strong>on</strong> of heart periods. It is alsopossible to compute the cardiac accelerati<strong>on</strong> (i.e. the derivative of the HR). Another methodc<strong>on</strong>sists of switching from the temporal to the frequency representati<strong>on</strong> of the HR signal. In thiscase, three frequency b<strong>and</strong>s are generally c<strong>on</strong>sidered:- the High Frequency (HF) b<strong>and</strong> including frequencies between 0.15Hz <strong>and</strong> 1Hz;- the Low Frequency (LF) b<strong>and</strong> ranging from 0.05Hz to 0.15Hz;- the Very Low Frequency (VLF) b<strong>and</strong> from 0.0033Hz to 0.05Hz.The exact boundaries of the different frequency b<strong>and</strong>s are still under discussi<strong>on</strong> in the literature<strong>and</strong> the given values are taken from [96]. The energy in the HF b<strong>and</strong> is known to be principallymediated by the parasympathetic activity while energy in the LF b<strong>and</strong> is influenced by the twoaut<strong>on</strong>omic systems. For those reas<strong>on</strong>s many studies have proposed to use the ratio of the LFenergy over the HF energy to reflect cardiac aut<strong>on</strong>omic balance [96]. Other methods are alsoavailable to quantify HRV such as determining HR accelerati<strong>on</strong> by computing the derivative ofthe HR signal.In order to compute HRV features it is first necessary to identify heart c<strong>on</strong>tracti<strong>on</strong>s to c<strong>on</strong>structthe IBI <strong>and</strong> HR time series. This can be d<strong>on</strong>e n<strong>on</strong>-invasively by analyzing the signals issued fromseveral devices. One soluti<strong>on</strong> is to record heart electrical activity by placing electrodes atappropriate positi<strong>on</strong>s [97]. The recording is called an electrocardiogram (ECG or EKG). From anECG it is possible to identify the peaks of the main waves, known as the R waves, whichcorresp<strong>on</strong>d to the main c<strong>on</strong>tracti<strong>on</strong>s of the heart. In order to measure HR, a microph<strong>on</strong>e can beused to m<strong>on</strong>itor heart sounds related to blood flow <strong>and</strong> valves activity of the heart. Finally, sinceheart pulses are visible in c<strong>on</strong>tinuous BP <strong>and</strong> BVP signals, it is also possible to compute HR fromthis type of signals [97].Increase or decrease of HR properties can be associated with different emoti<strong>on</strong>s [74]. Forinstance, Rainville et al. [6] observed an increase of mean HR <str<strong>on</strong>g>for</str<strong>on</strong>g> anger, fear, happiness <strong>and</strong>sadness compared to a neutral state. Moreover this increase was significantly different <str<strong>on</strong>g>for</str<strong>on</strong>g> fear37


Chapter 2<strong>and</strong> anger, fear <strong>and</strong> happiness as well as <str<strong>on</strong>g>for</str<strong>on</strong>g> fear <strong>and</strong> sadness. In the same study, cardiacaccelerati<strong>on</strong> was also a relevant variable <str<strong>on</strong>g>for</str<strong>on</strong>g> the discriminati<strong>on</strong> of those emoti<strong>on</strong>s while in [7] itwas shown to be significantly correlated with pleasantness. Ekman [5] found that HR was able toseparate happy, disgusted <strong>and</strong> surprised emoti<strong>on</strong>al states from angry, fearful <strong>and</strong> sad states.C<strong>on</strong>cerning HRV, an energy shift from HF to LF frequencies is known to be associated withparasympathetic withdrawal reacti<strong>on</strong>s [96]. Finally, a decrease of energy in the HF b<strong>and</strong>s wasobserved in [6] <str<strong>on</strong>g>for</str<strong>on</strong>g> fear <strong>and</strong> happiness compared to a neutral state while this energy wassignificantly different <str<strong>on</strong>g>for</str<strong>on</strong>g> fear <strong>and</strong> anger.d. Respirati<strong>on</strong>Respirati<strong>on</strong> involves organs such as lungs, airways <strong>and</strong> respiratory muscles. It is quantified byvariables like lung volume, tidal volume (the quantity of air moved during inhalati<strong>on</strong> <strong>and</strong>exhalati<strong>on</strong>), pressure, air flow <strong>and</strong> breathing rate which can be influenced by resistive properties<strong>and</strong> c<strong>on</strong>tracti<strong>on</strong> of the above-menti<strong>on</strong>ed organs [98]. Breathing is mostly achieved by thec<strong>on</strong>tracti<strong>on</strong> / relaxati<strong>on</strong> of the diaphragm <strong>and</strong> the intercostals muscles that increase / decrease thethoracic volume. C<strong>on</strong>trarily to other measures described above, the respirati<strong>on</strong> is operated byboth the ANS <strong>and</strong> the SNS since breathing is generally involuntary but it is possible to c<strong>on</strong>trol it<str<strong>on</strong>g>for</str<strong>on</strong>g> short periods of time. C<strong>on</strong>cerning the ANS divisi<strong>on</strong>s both the sympathetic <strong>and</strong>parasympathetic divisi<strong>on</strong>s are implied in respirati<strong>on</strong>.Precisely measuring respirati<strong>on</strong> characteristics requires the use of obtrusive equipment likespirometers <str<strong>on</strong>g>for</str<strong>on</strong>g> displaced air volume assessment <strong>and</strong> air-tight chambers <str<strong>on</strong>g>for</str<strong>on</strong>g> alveolar pressurecomputati<strong>on</strong> (the last <strong>on</strong>e being called body plethysmography). For a good review of suchapparatus the reader is referred to [98]. A less obtrusive device that detects respiratory patterns isthe respirati<strong>on</strong> belt. It is attached around the abdomen <strong>and</strong> / or the chest in order to measure theirexpansi<strong>on</strong> (Figure 2.11). Since this expansi<strong>on</strong> approximates the quantity of inspired <strong>and</strong> expiredair it is possible to measure the tidal volume with the proviso that calibrati<strong>on</strong> is per<str<strong>on</strong>g>for</str<strong>on</strong>g>medbe<str<strong>on</strong>g>for</str<strong>on</strong>g>eh<strong>and</strong> <strong>and</strong> that two belts are used <str<strong>on</strong>g>for</str<strong>on</strong>g> measuring both chest <strong>and</strong> abdomen extensi<strong>on</strong> [98, 99].Several types of belts exist, differing in the methods used to measure expansi<strong>on</strong> (inductivesensors, piezo-electric sensor, etc.). One disadvantage of the belts is that they are relativelysensitive to movements. Temperature sensors placed under the nostrils <strong>and</strong> be<str<strong>on</strong>g>for</str<strong>on</strong>g>e the mouth canalso be used to approximate respirati<strong>on</strong> flow. However this method is not really reliable,especially during fast inspirati<strong>on</strong> <strong>and</strong> expirati<strong>on</strong>s (Figure 2.11), prohibiting its usage <str<strong>on</strong>g>for</str<strong>on</strong>g> tidalvolume computati<strong>on</strong>.The main functi<strong>on</strong> of respirati<strong>on</strong> is to regulate the quantities of oxygen <strong>and</strong> carb<strong>on</strong> dioxidepresent in the blood to meet the needs of organs. However respirati<strong>on</strong> is also influenced byemoti<strong>on</strong>al processes [98, 99]. For instance, a low breathing rate is linked to relaxati<strong>on</strong> while38


State of the artirregular rhythm, quick variati<strong>on</strong>s, <strong>and</strong> cessati<strong>on</strong> of respirati<strong>on</strong> corresp<strong>on</strong>d to more arousedemoti<strong>on</strong>s like anger or fear [6, 100]. Laughing is also known to affect respirati<strong>on</strong> patterns, <str<strong>on</strong>g>for</str<strong>on</strong>g>instance by introducing high-frequency fluctuati<strong>on</strong>s in the signal measured by a respirati<strong>on</strong> belt.MovementartefactsNormalrespirati<strong>on</strong>Deep <strong>and</strong> slowrespirati<strong>on</strong>Norespirati<strong>on</strong>Fast <strong>and</strong> deeprespirati<strong>on</strong>timeDo notcorresp<strong>on</strong>d to realtidal volumetimeFigure 2.11. Examples of signals obtained from a respirati<strong>on</strong> belt tied across the chest <strong>and</strong> a temperaturesensor placed bellow the nostrils during different type of respirati<strong>on</strong>s.e. TemperatureThe internal temperature of the body is relatively stable while the temperature at the surface ofthe skin mainly depends <strong>on</strong> the surrounding temperature. The organism c<strong>on</strong>trols the internaltemperature by balancing heat producti<strong>on</strong> <strong>and</strong> heat loss. Heat producti<strong>on</strong> is achieved by musclec<strong>on</strong>tracti<strong>on</strong> (such as shivering), by increasing chemical <strong>and</strong> metabolic activity <strong>and</strong> byvasoc<strong>on</strong>stricti<strong>on</strong> of skin blood vessels. Heat loss is obtained through reducti<strong>on</strong> of metabolicactivity, vasodilatati<strong>on</strong> <strong>and</strong> sweating [101]. Since most of the heat is lost through the skin towardthe surrounding envir<strong>on</strong>ment, all those mechanisms also influence skin temperature.Skin temperature can be recorded by st<strong>and</strong>ard temperature sensors like thermometers. Analogthermometers which allow <str<strong>on</strong>g>for</str<strong>on</strong>g> digital recording of temperature are generally based <strong>on</strong> theelectrical resistance measurement of a piece of metal placed <strong>on</strong> the skin. To assess temperature,those devices make use of the predictable changes of the material resistance according to theheat. Less comm<strong>on</strong> devices include infrared thermal cameras. Since the amount of infrared lightemitted by a surface is proporti<strong>on</strong>al to its temperature it is possible to use such a device to39


Chapter 2m<strong>on</strong>itor thermal changes. The main advantage of this device is that it is not necessary to attach asensor <strong>on</strong> the skin, allowing <str<strong>on</strong>g>for</str<strong>on</strong>g> less obtrusive recordings.Since skin temperature is influenced by vasoc<strong>on</strong>stricti<strong>on</strong> <strong>and</strong> sweat, both being related toemoti<strong>on</strong>s as discussed in the preceding secti<strong>on</strong>s, it is not surprising that researchers found it to beassociated with emoti<strong>on</strong>al processes. Ekman [5] found a significant increase of skin temperature<str<strong>on</strong>g>for</str<strong>on</strong>g> anger compared to his five other basic emoti<strong>on</strong>s (sadness, happiness, fear, surprise <strong>and</strong>disgust). McFarl<strong>and</strong> [102] found that stimulating pers<strong>on</strong>s with emoti<strong>on</strong>al music led to an increaseof temperature <str<strong>on</strong>g>for</str<strong>on</strong>g> calm positive music <strong>and</strong> a decrease <str<strong>on</strong>g>for</str<strong>on</strong>g> excited negative pieces. In [103, 104]thermal images where investigated <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of stress <strong>and</strong> anxiety detecti<strong>on</strong>. Finally, sinceskin temperature is also related to muscular activati<strong>on</strong>s, a thermal camera could be used todistinguish between different facial expressi<strong>on</strong>s [105].402.2.3 Variability of physiological resp<strong>on</strong>sesAs described in the preceding secti<strong>on</strong>s, physiological signals are influenced by several functi<strong>on</strong>s<strong>and</strong> features of the organism. They are thus very sensitive to modificati<strong>on</strong>s of the internal statebut also to changes of the envir<strong>on</strong>ment. For instance, a change of the surrounding temperaturewill impact skin temperature directly, BVP because of vasoc<strong>on</strong>stricti<strong>on</strong> <strong>and</strong> GSR because ofpossible change in sudati<strong>on</strong>. A c<strong>on</strong>sequence is that the variability of physiological signals due tothe process under c<strong>on</strong>siderati<strong>on</strong>, in our case the emoti<strong>on</strong>, is often much lower than the variabilitydue to other phenomena. This last variability can then be c<strong>on</strong>sidered as an important disturbance<str<strong>on</strong>g>for</str<strong>on</strong>g> an emoti<strong>on</strong> assessment system.The variability is generally decomposed into two comp<strong>on</strong>ents: inter-participant variability <strong>and</strong>intra-participant variability. Inter-participant variability accounts <str<strong>on</strong>g>for</str<strong>on</strong>g> the differences inphysiological resp<strong>on</strong>ses from <strong>on</strong>e pers<strong>on</strong> to another. Those differences are due to several factorssuch as physiological characteristics, pers<strong>on</strong>al traits <strong>and</strong> behaviors. For instance body mass index,age <strong>and</strong> sex influence physiological resp<strong>on</strong>ses while smokers <strong>and</strong> n<strong>on</strong>-smokers may havedifferent respirati<strong>on</strong> patterns. Intra-participant variability accounts <str<strong>on</strong>g>for</str<strong>on</strong>g> the differences inphysiological resp<strong>on</strong>ses that can be observed <str<strong>on</strong>g>for</str<strong>on</strong>g> a given pers<strong>on</strong>. Those differences can beobserved from day to day <strong>and</strong> can also be due to sp<strong>on</strong>taneous events. Examples are a changes inmood from <strong>on</strong>e day to another that influences EEG waves [81], <strong>and</strong> coffee absorpti<strong>on</strong> thatmodifies many physiological resp<strong>on</strong>ses [91, 96, 97].It is thus important to find a way to reduce the impact of those noise sources. Intra-participantvariability can be reduced by having a high degree of c<strong>on</strong>trol over the experiment, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance by<str<strong>on</strong>g>for</str<strong>on</strong>g>bidding the ingesti<strong>on</strong> of coffee or any drug be<str<strong>on</strong>g>for</str<strong>on</strong>g>e physiological recordings <strong>and</strong> keeping trackof the mood of participants. It is worth to say that those methods can rarely be applied in the caseof an emoti<strong>on</strong> recogniti<strong>on</strong> system working in a real envir<strong>on</strong>ment, generally far away from any


State of the artexperimental c<strong>on</strong>trol. Picard [106] emphasized the importance of including baseline m<strong>on</strong>itoringto account <str<strong>on</strong>g>for</str<strong>on</strong>g> day to day variability in physiological signals. This method is equally important toreduce inter-participant variability. It c<strong>on</strong>sists in recording physiological signals during a“neutral” moment, be<str<strong>on</strong>g>for</str<strong>on</strong>g>e any emoti<strong>on</strong>al stimulati<strong>on</strong>. The signals or the features extracted fromthe signals during emoti<strong>on</strong> assessment are then expressed relatively to this baseline. Moreover, inemoti<strong>on</strong> assessment from physiological signals, normalizati<strong>on</strong> of the signals <str<strong>on</strong>g>for</str<strong>on</strong>g> each participantis often per<str<strong>on</strong>g>for</str<strong>on</strong>g>med to remove inter-participant variability [93, 106, 107]. Even if this method isquite effective, it is first required to have the range of all possible resp<strong>on</strong>ses to normalize thesignals which is generally not the case in real applicati<strong>on</strong>s.As shown in the preceding secti<strong>on</strong>s, different patterns of peripheral activati<strong>on</strong> were found <str<strong>on</strong>g>for</str<strong>on</strong>g> awide range of emoti<strong>on</strong>s. However, those findings are not always c<strong>on</strong>sistent since patterns whichare ascribed to <strong>on</strong>e emoti<strong>on</strong> tend to vary between studies [108]. One explanati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> thisphenomen<strong>on</strong> is that the physiological activati<strong>on</strong> that accompanies emoti<strong>on</strong>s is rather due to thetendency to acti<strong>on</strong> related to the emoti<strong>on</strong> than to the emoti<strong>on</strong> itself. In other words, the body ispreparing <str<strong>on</strong>g>for</str<strong>on</strong>g> acti<strong>on</strong> (running in case of fear, aggressive acti<strong>on</strong> in case of anger, etc.) <strong>and</strong> this isthe activity that can be measured <strong>on</strong> peripheral outputs. In [108] the authors proposed to unifythose two views by developing the c<strong>on</strong>text-deviati<strong>on</strong> specificity of emoti<strong>on</strong>s in which it isassumed that emoti<strong>on</strong>s have specific patterns of peripheral activati<strong>on</strong> in the case where thec<strong>on</strong>diti<strong>on</strong>s of emoti<strong>on</strong>al activati<strong>on</strong> are similar.These c<strong>on</strong>siderati<strong>on</strong>s are very important because they imply that any system dedicated to therecogniti<strong>on</strong> of affect should take into account the c<strong>on</strong>text specificity in its model of emoti<strong>on</strong>s.Though the goal of this thesis is not to design such a model, we believe that this c<strong>on</strong>cept shouldbe kept in mind when <strong>on</strong>e tries to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m affect recogniti<strong>on</strong>. From a practical point of view,taking into account all possible social, behavioral or psychological c<strong>on</strong>texts of emoti<strong>on</strong> elicitati<strong>on</strong>seems a very hard task. Fortunately, to greatly reduce the number of c<strong>on</strong>textual situati<strong>on</strong>s itgenerally suffices to c<strong>on</strong>centrate <strong>on</strong> <strong>on</strong>e particular applicati<strong>on</strong>. From this point of view, it seemsthat a good protocol <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> elicitati<strong>on</strong> is not the <strong>on</strong>e that will elicit emoti<strong>on</strong> as close aspossible to the real life emoti<strong>on</strong>s [107], but rather the <strong>on</strong>e that will elicit emoti<strong>on</strong>s underc<strong>on</strong>trolled or m<strong>on</strong>itored c<strong>on</strong>text close to the target applicati<strong>on</strong>.2.3 <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment from physiological signalsOver the last years, emoti<strong>on</strong> recogniti<strong>on</strong> from physiological signals has received much interest.The goal is to find a computati<strong>on</strong>al model that is able to associate a given physiological pattern toan emoti<strong>on</strong>al state. This can be d<strong>on</strong>e by several methods discussed in this chapter.Table 2.4 provides a (n<strong>on</strong> exhaustive) list of relevant studies c<strong>on</strong>cerning emoti<strong>on</strong> assessmentfrom physiological signals. Un<str<strong>on</strong>g>for</str<strong>on</strong>g>tunately, it is difficult to make comparis<strong>on</strong>s between these41


Chapter 2studies because they differ <strong>on</strong> several criteria. Six criteria are introduced below to help the readergain insight into the current state of the art as well as to discuss important aspects of emoti<strong>on</strong>assessment from physiological signals. The studies are listed in Table 2.4 according to those sixcriteria.Number of participants: in a study that includes a high number of participants the results can beregarded as more significant. Another point of importance is to know if the model obtained <str<strong>on</strong>g>for</str<strong>on</strong>g>emoti<strong>on</strong> identificati<strong>on</strong> is user-specific or not. A user specific model avoids the problems relatedto inter-participant variability but a new model will have to be generated <str<strong>on</strong>g>for</str<strong>on</strong>g> each new user.<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> elicitati<strong>on</strong>: how to elicit emoti<strong>on</strong>s from a participant is also a questi<strong>on</strong> of significantinterest, especially c<strong>on</strong>sidering the importance of the c<strong>on</strong>text as detailed be<str<strong>on</strong>g>for</str<strong>on</strong>g>e. Picard [106]proposed five factors to describe the c<strong>on</strong>text in which the emoti<strong>on</strong> are elicited. One of those fivefactors divides emoti<strong>on</strong> elicitati<strong>on</strong> approaches into two categories: subject-elicited <strong>and</strong> eventelicited.In the first category, emoti<strong>on</strong>s can be generated by asking the participant to act as ifhe/she was feeling a particular emoti<strong>on</strong> (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance by mimicking the facial expressi<strong>on</strong> of anger)or to remember a past emoti<strong>on</strong>al event of his / her life. This method has often been used in facialexpressi<strong>on</strong> recogniti<strong>on</strong> <strong>and</strong> it has been shown in [5] that it is effective to induce specificperipheral activity. In the sec<strong>on</strong>d category, it is possible to use images, sounds, video clips or anyemoti<strong>on</strong>ally evocative task. Several databases of stimuli have been designed <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose ofemoti<strong>on</strong> elicitati<strong>on</strong> like the Internati<strong>on</strong>al <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Picture System or Internati<strong>on</strong>al DigitizedSound system (IAPS, IADS)[54]. These databases are generally accompanied by affectiveevaluati<strong>on</strong>s from experts or average judgments of several people. However, since past experienceplays a key role in emoti<strong>on</strong> elicitati<strong>on</strong>, it can also be important to ask the user to report <strong>and</strong> selfassesshis / her feelings. <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> elicitati<strong>on</strong> is also influenced by the number <strong>and</strong> the complexityof the targeted emoti<strong>on</strong>s.Time: temporal aspects are also relevant <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> recogniti<strong>on</strong> but have <strong>on</strong>ly received littleattenti<strong>on</strong>. According to [20], it is possible to define some time categories that range from the “fullblown emoti<strong>on</strong>s”, lasting <str<strong>on</strong>g>for</str<strong>on</strong>g> some sec<strong>on</strong>ds or minutes, to the traits, lasting <str<strong>on</strong>g>for</str<strong>on</strong>g> years if not all thelifetime. In between are categories such as moods or emoti<strong>on</strong>al disorders. In human computerinteracti<strong>on</strong>, most of the applicati<strong>on</strong>s under c<strong>on</strong>siderati<strong>on</strong> deal with what Cowie defines as “fullblown emoti<strong>on</strong>s” thus managing phenomena that last from sec<strong>on</strong>ds to minutes. In an idealaffective interface, the emoti<strong>on</strong> of the user should be detected as so<strong>on</strong> as possible (let’s say in fewsec<strong>on</strong>ds) in order to take the proper decisi<strong>on</strong> that directly matches the user expectati<strong>on</strong> <strong>and</strong> not<strong>on</strong>e that was expected minutes be<str<strong>on</strong>g>for</str<strong>on</strong>g>e. Synchr<strong>on</strong>izati<strong>on</strong> of the different modalities is also an issuesince the activati<strong>on</strong> of physiological outputs can occur at different time scales. For instance, achange in temperature of the skin is much slower than a change in brain activity occurring a fewmillisec<strong>on</strong>ds after emoti<strong>on</strong> elicitati<strong>on</strong>.42


State of the artRef.Kim J.[100]Lisetti etal.[93]Rainvilleet al.[6]Picard etal.[106]Katsis etal.[109]Number ofparticipants3user specific29not userspecific43not userspecific 6Elicitati<strong>on</strong>methodsMusic,AuDB(Augsburgerdatabase ofbio-signals)TemporalaspectsTime of atrial notspecified25recordingsover 25daysFilm clips 70-231 s GSR, heart rate,temperatureSelfInducti<strong>on</strong>1 Selfinducti<strong>on</strong>10not userspecificDrivingsimulati<strong>on</strong>Signals / sensors <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> classes Methods (classifiers,feature selecti<strong>on</strong>, etc.)Skin c<strong>on</strong>ductance, Joy, anger, relaxati<strong>on</strong>, SFFS feature selecti<strong>on</strong>,EMG, Respirati<strong>on</strong>, sadnessLDA with MSEECGPositive / negative SFFS feature selecti<strong>on</strong>,LDA with MSE90 s ECG, respirati<strong>on</strong>,skin c<strong>on</strong>ductance,EMG (zygomatic,masseter,corrugator)100 to250 s20differentdays ofrecordingFeaturescomputed<strong>on</strong> 10 swindowsEMG, GSR,respirati<strong>on</strong>, BVP,ECGEMG, ECG,respirati<strong>on</strong>, GSRHigh / low arousalSadness, amusement,fear, anger, frustrati<strong>on</strong>,surpriseAnger (15 part.), fear(15 part.), happiness (15part.), sadness (17 part.) 6Neutral (no-emoti<strong>on</strong>),anger, hate, grief,plat<strong>on</strong>ic love, romanticlove, joy, reverenceHigh / low arousalPositive / negativeHigh stress, low stress,disappointment <strong>and</strong>euphoriaSFFS feature selecti<strong>on</strong>,LDA with MSENeural network withMarquardt backpropagati<strong>on</strong>Step wise discriminantanalysisBestresults84%84%94%84%49%SFFS-Fisher projecti<strong>on</strong> 81%SVMAdaptive neuro-fuzzyinference system84%87%79%77%6 The physiological activity of each participant was recorded <str<strong>on</strong>g>for</str<strong>on</strong>g> <strong>on</strong>ly <strong>on</strong>e or two of the emoti<strong>on</strong>al c<strong>on</strong>diti<strong>on</strong>s. The final number of participant <str<strong>on</strong>g>for</str<strong>on</strong>g> each c<strong>on</strong>diti<strong>on</strong> is given inthe “<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> classes” column.43


Chapter 244Ref.Kim K.H.et al.[110]Wagner etal.[111]Haag etal.[107]Sibha etal.[112]Takahashiet al.[113]Le<strong>on</strong> etal.[114]Number of Elicitati<strong>on</strong>participants methods50 children Combinati<strong>on</strong>of storynot user telling,specific visualisati<strong>on</strong><strong>and</strong> audiostimulus1 Musicchosen bytheparticipant1 images fromIAPS27not userspecific12not userspecific9not userspecificSelfinducti<strong>on</strong>Film clipsimages fromthe IAPSTemporalaspectsSignals / sensors <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> classes Methods (classifiers,feature selecti<strong>on</strong>, etc.)50 s Skin temperature, Sadness, anger, stress Subjects <str<strong>on</strong>g>for</str<strong>on</strong>g> training <strong>and</strong>GSR, heart rateothers <str<strong>on</strong>g>for</str<strong>on</strong>g> testingSadness, anger, stress, SVMsurprise2 min25recordingsover 25days2 sseveraldays60 s2recordingsessi<strong>on</strong>s<strong>on</strong>differentdaysTime of atrial notEMG, ECG, GSR,respirati<strong>on</strong>EMG, GSR, Skintemperature, BVP,ECG, respirati<strong>on</strong>ECG, GSR, fingertemperature, bloodpressure, EOG,EMG (zygomatic,corrugator,masseter, depressormuscles)EEG, BVP, GSRspecified6 s Heart rate, GSR,BVPAnger, sadness, joy,pleasurePositive / negativeHigh / low arousalArousalValenceFear, anger, neutralJoy, anger, sadness, fear,relaxati<strong>on</strong>Neutral, negative,positiveLDA, KNN, MLPSFFS, Fisher, ANOVANeural network <str<strong>on</strong>g>for</str<strong>on</strong>g>regressi<strong>on</strong>Accury is computed as thenumber of samples that fallin a 20% b<strong>and</strong>width of thecorrect valueLDA (first sessi<strong>on</strong> astraining set, sec<strong>on</strong>d as testset)Bestresults78%62%92%86%96%97%90%67%Linear SVM <strong>on</strong>e vs. all 42%Autoassociative neuralnetworks1 participant <str<strong>on</strong>g>for</str<strong>on</strong>g> testingothers <str<strong>on</strong>g>for</str<strong>on</strong>g> training71%


Ref.Sakata etal.[115]Rani et al.[116]State of the artNumber of Elicitati<strong>on</strong> Temporal Signals / sensors <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> classes Methods (classifiers, Bestparticipants methods aspectsfeature selecti<strong>on</strong>, etc.) results16 Pictures 3 s EEG6 emoti<strong>on</strong>s LDA 29%15user specificsolvinganagrams,playingp<strong>on</strong>g3-4 min6 sessi<strong>on</strong>s<strong>on</strong>differentdaysHear rate (resultsnot presented)ECG, GSR, bioimpedance,EMG(corrugator,zygomatic,trapezius), temp.,BVP, heart sound3 levels of intensity <str<strong>on</strong>g>for</str<strong>on</strong>g>:engagement, anxiety,boredom, frustrati<strong>on</strong>,anger.A classifier is trainedindependently <str<strong>on</strong>g>for</str<strong>on</strong>g> eachemoti<strong>on</strong>. Results are theaverage accuracy acrossparticipants <strong>and</strong>affective statesKNNRegressi<strong>on</strong> treeBayes networkTable 2.4. List of publicati<strong>on</strong>s <strong>on</strong> emoti<strong>on</strong> assessment from physiological signals. Signals acr<strong>on</strong>yms are: Electromyography (EMG), Electrocardiogram (ECG),Galvanic Skin Resp<strong>on</strong>se (GSR), Electroencephalography (EEG), Blood Volume Pulse (BVP). Classificati<strong>on</strong> acr<strong>on</strong>yms are : Sequential Floating Forward Search(SFFS), Linear discriminant analysis (LDA), Support Vector Machine (SVM), Mean Square Error (MSE), Multi Layer Perceptr<strong>on</strong> (MLP), K-Nearest Neighbors(KNN), ANalysis Of Variance (ANOVA).SVM79%83%78%86%45


Chapter 2Sensors / modalities: as described in chapter 2.2 various sensors can be used to measurephysiological activity related to emoti<strong>on</strong>al processes. However, sensors used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>assessment should be chosen carefully so that they do not perturb the user. First, sensors shouldnot be uncom<str<strong>on</strong>g>for</str<strong>on</strong>g>table <str<strong>on</strong>g>for</str<strong>on</strong>g> the user in order to avoid parasite emoti<strong>on</strong>s like pain. Sec<strong>on</strong>dly, theyshould not prevent the use of classical modalities <str<strong>on</strong>g>for</str<strong>on</strong>g> instance by m<strong>on</strong>opolizing the h<strong>and</strong>s of theuser. When physiological sensors are used <str<strong>on</strong>g>for</str<strong>on</strong>g> affective computing they switch from the st<strong>and</strong>ardstatus of sensors to the c<strong>on</strong>cept of modalities that communicate in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> about the emoti<strong>on</strong>alstatus of the user. It is then necessary to merge these modalities to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m emoti<strong>on</strong>s assessmentin a reliable way, taking into account redundant <strong>and</strong> complementary in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> such as therelati<strong>on</strong> that exists between HRV <strong>and</strong> respirati<strong>on</strong> [117, 118].<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> models / classes: as discussed in chapter 2.1, several representati<strong>on</strong>s of emoti<strong>on</strong>s areavailable. Secti<strong>on</strong> 2.1.4 discusses their usability <strong>and</strong> extensi<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of emoti<strong>on</strong>assessment. In order to define a computati<strong>on</strong>al model that links physiological reacti<strong>on</strong>s toaffective states, it is necessary to record physiological activity together with a ground-truth (i.e.the true elicited emoti<strong>on</strong>al state). However, c<strong>on</strong>structing an emoti<strong>on</strong>al ground-truth is a difficulttask because emoti<strong>on</strong>al states are influenced by several comp<strong>on</strong>ents of the organism as explainedin Secti<strong>on</strong> 2.1.3.b. Several methods can be used to c<strong>on</strong>struct a ground-truth, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance bydefining the emoti<strong>on</strong>al state a-priori (generally based <strong>on</strong> precedent studies or evaluati<strong>on</strong>s of thestimuli) or by asking the participants to self-report their feelings. Depending <strong>on</strong> the methodsused, the ground-truth can be quite different. The advantages <strong>and</strong> disadvantages of the groundtruthc<strong>on</strong>structi<strong>on</strong> methods are discussed in Secti<strong>on</strong> 4.1.1.Methods: a wide range of methods has been used to infer affective states. Most of them are partof the machine learning <strong>and</strong> pattern recogniti<strong>on</strong> techniques. Classifiers like k-Nearest Neighbors(KNN), Linear Discriminant Analysis (LDA), neural networks, Support Vector Machines(SVM’s) <strong>and</strong> others [119, 120] are useful to detect emoti<strong>on</strong>al classes of interest. Regressi<strong>on</strong>techniques [120] can also be used to obtain c<strong>on</strong>tinuous estimati<strong>on</strong> of emoti<strong>on</strong>s, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance in thevalence-arousal space. Prior to inferring emoti<strong>on</strong>al states it is important to define somephysiological features of interest. It is very challenging to find with certainty some features inphysiological signals that always correlate with the affective status of users. Those variablesfrequently differ from <strong>on</strong>e user to another <strong>and</strong> they are also very sensitive to day to day variati<strong>on</strong>sas well as to the c<strong>on</strong>text of the emoti<strong>on</strong> inducti<strong>on</strong>. To per<str<strong>on</strong>g>for</str<strong>on</strong>g>m this selecti<strong>on</strong> researchers generallyapply feature selecti<strong>on</strong> or projecti<strong>on</strong> algorithms like Sequential Floating Forward Search (SFFS)or Fisher projecti<strong>on</strong>.As can be seen from Table 2.4 there are large differences in classificati<strong>on</strong> accuracy even <str<strong>on</strong>g>for</str<strong>on</strong>g>studies that employ the same number of classes. Although this can be partly explained by thefactors detailed above, we believe that such variati<strong>on</strong>s mostly result from: the differences in46


State of the artemoti<strong>on</strong> elicitati<strong>on</strong> strategies, the type of physiological signals (modalities) used, <strong>and</strong> the chosenmodel of emoti<strong>on</strong>s (or emoti<strong>on</strong>al classes).For emoti<strong>on</strong>s elicited by an event, the best results obtained <strong>on</strong> more than three classes are thosepresented in [93] <strong>and</strong> [111]. One interesting point is that even though these studies use differentstimuli (film clip versus music) they both use a method to c<strong>on</strong>trol <str<strong>on</strong>g>for</str<strong>on</strong>g> the validity of the elicitedemoti<strong>on</strong>s. In [93], the film clips presented were chosen according to the results obtained in a pilotstudy. In this pilot study, several participants were asked to evaluate clips by stating the emoti<strong>on</strong>they felt as well as its intensity. In [111], the authors worked with the Augsburger Database ofBiosignals (AuDB), the participants where asked to freely choose music that matched the desiredemoti<strong>on</strong>s. Both of these methods insure that emoti<strong>on</strong>s felt during the experiment are intense <strong>and</strong>corresp<strong>on</strong>d to the expected <strong>on</strong>es. The good results obtained in [106] using a self inducti<strong>on</strong>protocol also tend to c<strong>on</strong>firm the importance of reliable elicitati<strong>on</strong>. In [106], the participant wasthe experimenter herself so that the emoti<strong>on</strong>s to be induced were perfectly clear to her. Moreover,the participant remembered past emoti<strong>on</strong>al events <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> elicitati<strong>on</strong>; this implies a str<strong>on</strong>gintensity since remembered events are generally those that have induced intense feelings.Diverse types of physiological activity measurements from both the peripheral <strong>and</strong> the centralnervous system have been used. Up to now, most of the studies using brain activity <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>assessment have used EEG signals with unc<strong>on</strong>vincing results [36, 115]. One could c<strong>on</strong>clude thatEEG signals in the present state of the art are not effective <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment; the presentthesis however will argue against this. Describing the state of the art <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> recogniti<strong>on</strong>based <strong>on</strong> peripheral signals is a real challenge because most studies per<str<strong>on</strong>g>for</str<strong>on</strong>g>med classificati<strong>on</strong> <strong>on</strong>feature sets that include features from many types of signals, thus preventing analysis <str<strong>on</strong>g>for</str<strong>on</strong>g> singlemodalities. However, there are signals that are employed in nearly all studies, like GSR <strong>and</strong> HR(extracted from ECG or plethysmography). These two signals are known to correlate well withaffective states [5, 7, 92]. In many results it can also be seen that EMG signals are relevant <str<strong>on</strong>g>for</str<strong>on</strong>g>emoti<strong>on</strong> assessment [6, 111, 112, 116, 121]. Generally, EMG electrodes are positi<strong>on</strong>ed tomeasure facial muscle activity. Muscles that are often recorded include: the venter fr<strong>on</strong>talis(raising of eyebrows), the zygomatic (smiling) <strong>and</strong> the corrugator supercili (frowning ofeyebrows). Since emoti<strong>on</strong>al states are often associated with facial expressi<strong>on</strong>s, measuring facialactivity is str<strong>on</strong>gly relevant to assess emoti<strong>on</strong>s. However having EMG, electrodes <strong>on</strong> the face isquite uncom<str<strong>on</strong>g>for</str<strong>on</strong>g>table which could hamper their usage in a c<strong>on</strong>crete applicati<strong>on</strong>.In order to assess emoti<strong>on</strong>s using physiological signals it is necessary to extract features fromthose signals that are relevant to the problem. In the current state of the art, most of studies [93,106, 107, 122] c<strong>on</strong>sider physiological signals as independent Gaussian processes <strong>and</strong> thus extractstatistical features such as mean, st<strong>and</strong>ard deviati<strong>on</strong> <strong>and</strong> sometimes moments of higher order(skewness <strong>and</strong> kurtosis). However, it is likely that this independency assumpti<strong>on</strong> does not hold47


Chapter 2since physiological signals are influenced by identical systems such as the ANS <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheralsignals. Moreover, correlati<strong>on</strong>s exist between the different signals because they are modulated bya single reacti<strong>on</strong> to emoti<strong>on</strong>s. For instance both finger blood pressure <strong>and</strong> skin temperature areinfluenced by vasoc<strong>on</strong>stricti<strong>on</strong>. To our knowledge, there are no studies that try to c<strong>on</strong>sider thoseinteracti<strong>on</strong>s at the feature extracti<strong>on</strong> level. Most of the studies per<str<strong>on</strong>g>for</str<strong>on</strong>g>m feature selecti<strong>on</strong> afterfeature extracti<strong>on</strong> with the goal of removing redundant features <strong>and</strong> keeping relevant <strong>and</strong>synergetic <strong>on</strong>es. Other features frequently extracted in the temporal domain are statistical features(mean, st<strong>and</strong>ard deviati<strong>on</strong>, etc.) of the first <strong>and</strong> sec<strong>on</strong>d derivative of the signal [87, 93, 106, 113,123]. Apart from temporal features, features from the frequency domain are also frequentlyextracted especially <str<strong>on</strong>g>for</str<strong>on</strong>g> HR, respirati<strong>on</strong> <strong>and</strong> EEG [6, 109, 110, 115, 122, 124].One of the most obvious observati<strong>on</strong>s that can be made from Table 2.4 is that different models ofemoti<strong>on</strong>s lead to different classificati<strong>on</strong> accuracies. This is especially clear when comparing thebasic-emoti<strong>on</strong>s models, which generally include more than three categories, to emoti<strong>on</strong>s in thevalence-arousal space model, including two or three categories. Thanks to the works of Wagner[111] <strong>and</strong> Picard [106], it is possible to compare results <strong>on</strong> valence-arousal classes to thoseobtained <strong>on</strong> basic emoti<strong>on</strong>s classes in an intra-study framework. The c<strong>on</strong>clusi<strong>on</strong> emerging fromthis comparis<strong>on</strong> is that the accuracies reported with the valence-arousal representati<strong>on</strong> are lowerthan the accuracies reported with basic emoti<strong>on</strong>s, c<strong>on</strong>sidering that the number of classes isdifferent. A possible explanati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> this is that acquisiti<strong>on</strong> protocols were designed <str<strong>on</strong>g>for</str<strong>on</strong>g> basicemoti<strong>on</strong> elicitati<strong>on</strong> while valence-arousal classes were defined by grouping basic emoti<strong>on</strong>s intotwo sets of classes (high vs. low arousal <strong>and</strong> high vs. low valence). This coarse definiti<strong>on</strong> canlead to c<strong>on</strong>fusing classes. For instance, grouping bliss <strong>and</strong> joy in the same positive class is pr<strong>on</strong>eto errors since peripheral activity <str<strong>on</strong>g>for</str<strong>on</strong>g> bliss, which is a rather calm feeling, is certainly different tothe <strong>on</strong>e of excited joy. However, results from other studies indicate that classificati<strong>on</strong> in thevalence-arousal space is often associated with lower accuracies than when using basic emoti<strong>on</strong>labels. Moreover, identificati<strong>on</strong> of valence classes is generally harder than identificati<strong>on</strong> ofarousal classes (Table 2.4), which supports the idea that peripheral activity has a highercorrelati<strong>on</strong> with arousal than valence [7]. No clear differences in accuracy can be observedbetween the studies using different labels <strong>and</strong> number of classes to recognize basic emoti<strong>on</strong>s.Only some of the studies listed in Table 2.4 per<str<strong>on</strong>g>for</str<strong>on</strong>g>med emoti<strong>on</strong> recogniti<strong>on</strong> <strong>on</strong>line [25, 114, 125].The others per<str<strong>on</strong>g>for</str<strong>on</strong>g>med acquisiti<strong>on</strong> of the signals in a well c<strong>on</strong>trolled experimental envir<strong>on</strong>ment<strong>and</strong>, in a sec<strong>on</strong>d step, applied emoti<strong>on</strong> recogniti<strong>on</strong> algorithms offline making use of crossvalidati<strong>on</strong>methods to determine the effectiveness of their system <strong>on</strong> unseen data. In [114] theauthors first trained a model to associate physiological patterns to emoti<strong>on</strong>s <strong>and</strong> then applied thesame model to detect emoti<strong>on</strong>s during the presentati<strong>on</strong> of images from the IAPS. As can be seenfrom Table 2.4 the obtained accuracy <strong>on</strong> three classes (neutral, negative <strong>and</strong> positive) is quite48


State of the arthigh (71%) dem<strong>on</strong>strating the feasibility of <strong>on</strong>line emoti<strong>on</strong> recogniti<strong>on</strong>. However, <strong>on</strong>ly <strong>on</strong>eparticipant was tested with <strong>on</strong>line detecti<strong>on</strong> preventing further c<strong>on</strong>clusi<strong>on</strong>s c<strong>on</strong>cerning thereproducibility of this approach. Moreover, recordings <strong>and</strong> <strong>on</strong>line classificati<strong>on</strong> where alsoper<str<strong>on</strong>g>for</str<strong>on</strong>g>med in experimental c<strong>on</strong>diti<strong>on</strong>s that are far from ecological <strong>and</strong> real applicati<strong>on</strong> c<strong>on</strong>diti<strong>on</strong>s.Another example of <strong>on</strong>line emoti<strong>on</strong> assessment is [125] where the authors designed a fuzzyemoti<strong>on</strong> detecti<strong>on</strong> system <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of studying emoti<strong>on</strong>s during game play. However, no<strong>on</strong>line accuracy is reported <str<strong>on</strong>g>for</str<strong>on</strong>g> their system. Instead, the authors directly use the assessedemoti<strong>on</strong>s to compare different game play c<strong>on</strong>diti<strong>on</strong>s. One of the most relevant studies c<strong>on</strong>cerning<strong>on</strong>line emoti<strong>on</strong> recogniti<strong>on</strong> is certainly [25] where pleasure (low vs. high) is assessed from thephysiological signals of autistic children in order to adapt the behavior of a robot arm playingwith them. The objective is to c<strong>on</strong>struct a robot able to change its gaming strategies as apsychologist would do. In this case the authors reported an average <strong>on</strong>line accuracy of 81.1%when the two assessed emoti<strong>on</strong>al classes where compared to a ground-truth based <strong>on</strong>psychologist judgments. In this study, emoti<strong>on</strong>s were assessed while the children were interactingwith the robot thus showing the possibility of emoti<strong>on</strong> assessment <str<strong>on</strong>g>for</str<strong>on</strong>g> applicati<strong>on</strong>s in a real HMIenvir<strong>on</strong>ment.49


Physiological signals recording <strong>and</strong> processingChapter 3Physiological signals recording <strong>and</strong> processing3.1 MaterialThe material used <str<strong>on</strong>g>for</str<strong>on</strong>g> the acquisiti<strong>on</strong> of physiological signals is the Biosemi Active 2 System[126]. This system is composed of several comp<strong>on</strong>ents as detailed in Figure 3.1. The Biosemisystem allows <str<strong>on</strong>g>for</str<strong>on</strong>g> simultaneous acquisiti<strong>on</strong> of EEG <strong>and</strong> peripheral signals. The analog signals aretransmitted to the A/D c<strong>on</strong>verter <str<strong>on</strong>g>for</str<strong>on</strong>g> amplificati<strong>on</strong> <strong>and</strong> c<strong>on</strong>versi<strong>on</strong> to digital data. The data is thentransmitted to the receiver through an optical fiber <strong>and</strong> finally to the PC <str<strong>on</strong>g>for</str<strong>on</strong>g> visualizati<strong>on</strong> <strong>and</strong>storage, both operated by the Actiview software (v. 5.21). Since the A/D c<strong>on</strong>verter is suppliedwith a low power battery <strong>and</strong> galvanic insulati<strong>on</strong> from the main power supply is ensured by theoptical fiber, there is no risk of electrocuti<strong>on</strong>. The system also provides a way to plug in externaltriggers that will be recorded with the other data sources. This is useful <str<strong>on</strong>g>for</str<strong>on</strong>g> synchr<strong>on</strong>izati<strong>on</strong> withanother acquisiti<strong>on</strong> system <strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> recording of manual event triggers (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance by pushingbutt<strong>on</strong>s).EEGA / Dc<strong>on</strong>verterUSBreceiverActiviewsoftwarePCPeriph.signalsOpticalfiberExternaltriggersUSBSoft.triggersProtocolprogramHard driveFigure 3.1. Hardware <strong>and</strong> software <str<strong>on</strong>g>for</str<strong>on</strong>g> signal acquisiti<strong>on</strong>.During <strong>and</strong> after the recording of physiological data it is necessary to have a program thatpresents stimuli <strong>and</strong> records participant answers to electr<strong>on</strong>ic questi<strong>on</strong>naires. In Figure 3.1 this isnamed the “Protocol program”. The protocol describes the scheduling of the different tasks thathave to be per<str<strong>on</strong>g>for</str<strong>on</strong>g>med by the participants during the experiment. It is important to synchr<strong>on</strong>ize thestimuli presented by the protocol program with the recorded signals to be able to retrieve thephysiological signals that corresp<strong>on</strong>d to the events of interest. Since the Actiview software isimplemented in Labview, the same plat<str<strong>on</strong>g>for</str<strong>on</strong>g>m was used to develop the protocol programs. In thisway, we were able to send software triggers <str<strong>on</strong>g>for</str<strong>on</strong>g> the different events that were then recorded bythe Actiview software together with physiological data. However, since the operating system wasnot real time there was nothing that could guarantee the precisi<strong>on</strong> of the software triggers.Experimentati<strong>on</strong> showed that the lag between the stimulus presentati<strong>on</strong> <strong>and</strong> the recorded triggerdid not exceed 100 ms.51


Chapter 3All the data were recorded in a room relatively immune to electromagnetic noise or in the EckelC14 audiometric research chamber with electromagnetic insulati<strong>on</strong> 7 (2.16m x 1.80m x 2.37m). Inboth cases aerati<strong>on</strong> ensured that the ambient temperature did not increase due to the presence of aparticipant. The Faraday cage also provided acoustic isolati<strong>on</strong> to ensure that participants were notdisturbed by audio noise.3.1.1 EEGIn order to avoid invasive recordings, surface EEG electrodes were used to m<strong>on</strong>itor brain activity.The Biosemi system uses Ag-AgCl active electrodes to record electro potentials at the surface ofthe scalp which implies that a small current is applied <strong>on</strong> the skin. This design allows <str<strong>on</strong>g>for</str<strong>on</strong>g> noise<strong>and</strong> impedance reducti<strong>on</strong>, low offset voltages <strong>and</strong> stable DC per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance [126]. Two electrodes,the CMS <strong>and</strong> DRL electrodes, replace the usual ground electrode comm<strong>on</strong> to other EEGacquisiti<strong>on</strong> systems. As can be seen from Figure 3.2 which gives an overview of the names <strong>and</strong>positi<strong>on</strong>s of the EEG electrodes, the CMS <strong>and</strong> DRL electrodes were placed respectively <strong>on</strong> theleft <strong>and</strong> right side of POz electrode positi<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> all the recordings. The raw signals are referencedto the CMS electrode but it is necessary to re-reference them to obtain a better signal to noiseratio [126] (see Secti<strong>on</strong> 3.3.2).In the different experiments per<str<strong>on</strong>g>for</str<strong>on</strong>g>med, two electrode c<strong>on</strong>figurati<strong>on</strong>s were employed <str<strong>on</strong>g>for</str<strong>on</strong>g> EEGrecordings. In the first c<strong>on</strong>figurati<strong>on</strong> 64 electrodes were used to record an EEG with a relativelyhigh spatial resoluti<strong>on</strong> (Figure 3.2). Those electrodes were positi<strong>on</strong>ed according the 10-10 system[76] which proposes positi<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> up to 74 electrodes. The sec<strong>on</strong>d c<strong>on</strong>figurati<strong>on</strong> c<strong>on</strong>sists of 19electrodes placed according to the 10-20 system which allows <str<strong>on</strong>g>for</str<strong>on</strong>g> up to 21 positi<strong>on</strong>s [76]. Thechosen electrodes are colored in green in Figure 3.2. The sec<strong>on</strong>d c<strong>on</strong>figurati<strong>on</strong> permits to reducethe time required to plug electrodes into the headcap. It is then possible to acquire data from moreparticipants in a given amount of time, the main drawback being the loss of spatial resoluti<strong>on</strong>.As can be seen from Figure 3.2 electrodes are plugged in electrode holders, themselves fixed <strong>on</strong> acap attached <strong>on</strong> the head of participants. A cap is first chosen according to the head size of eachparticipant <strong>and</strong> fitted <strong>on</strong> the head so that:- Cz, Fpz <strong>and</strong> Oz electrodes are <strong>on</strong> the nasi<strong>on</strong>-ini<strong>on</strong> line (Figure 3.2),- Fpz is at approximately 10% of the nasi<strong>on</strong>-ini<strong>on</strong> distance from the nasi<strong>on</strong>,- Oz is at approximately 10% of the nasi<strong>on</strong>-ini<strong>on</strong> distance from the ini<strong>on</strong>.7 http://www.eckel.ca/ (retreived <strong>on</strong> 29 April 2009)52


Physiological signals recording <strong>and</strong> processingDuring the experiment, the cap was attached using Velcro straps to prevent displacement ofelectrodes. Be<str<strong>on</strong>g>for</str<strong>on</strong>g>e plugging an electrode, it is m<strong>and</strong>atory to fill in the corresp<strong>on</strong>ding electrodeholder with a salty gel ensuring c<strong>on</strong>tact <strong>and</strong> c<strong>on</strong>ductivity between electrodes <strong>and</strong> the skin. The gelmust be inserted using a syringe because of the small size of the electrode holders. Attenti<strong>on</strong>should be paid not to insert too much gel into the holders so that two different electrodes wouldbe c<strong>on</strong>nected by a gel bridge <strong>and</strong> record the same broad activity. This is particularly important <str<strong>on</strong>g>for</str<strong>on</strong>g>the CMS / DRL electrodes since they are close to each other <strong>and</strong> a direct c<strong>on</strong>necti<strong>on</strong> of thoseelectrodes would reduce the signal to noise ratio <str<strong>on</strong>g>for</str<strong>on</strong>g> all electrodes. Two types of indicators wereused to c<strong>on</strong>trol that the electrodes were correctly plugged: impedance of the electrodes was keptbelow 5 K <strong>and</strong> signals were visually verified. Visual verificati<strong>on</strong> c<strong>on</strong>sisted in ensuring thatthere is no sudden drift in the signals, c<strong>on</strong>trolling the appearance of alpha waves when theparticipant relaxes with eyes closed <strong>and</strong> observing the appropriate shape of the signals during eyeblinks. Setting up the cap with 64 electrodes takes approximately 30 minutes <str<strong>on</strong>g>for</str<strong>on</strong>g> a trainedexperimenter.Nasi<strong>on</strong>Ini<strong>on</strong>Figure 3.2. (left) A participant wearing the EEG cap with 64 electrodes plugged. (right) Top head view withthe positi<strong>on</strong>s <strong>and</strong> names of the 64 electrodes used <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG recording. For a 19 electrodes c<strong>on</strong>figurati<strong>on</strong> <strong>on</strong>lythe green electrodes were used.3.1.2 Peripheral sensorsThe sensors used to m<strong>on</strong>itor peripheral activity during emoti<strong>on</strong>al stimulati<strong>on</strong>s are presented inFigure 3.3. Apart from the respirati<strong>on</strong> belt all the other sensors were attached <strong>on</strong> <strong>on</strong>e of the h<strong>and</strong>s.More precisely, they were placed <strong>on</strong> the left (resp. right) h<strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> right-h<strong>and</strong>ed (resp. left-h<strong>and</strong>ed)people to leave their preferred h<strong>and</strong> free to interact with the computer <strong>and</strong> answer questi<strong>on</strong>naires.53


Chapter 3A GSR (Galvanic Skin Resp<strong>on</strong>se) sensor, which measures the resistance of the skin in anexosomatic setting (see Secti<strong>on</strong> 2.2.2.a), was used to record EDA. This sensor is composed oftwo Ag-AgCl electrodes with a diameter of 8 mm that were placed <strong>on</strong> the distal phalanges of theindex <strong>and</strong> middle fingers (Figure 3.3) as suggested by the guidelines of the society <str<strong>on</strong>g>for</str<strong>on</strong>g>psychophysiological research [127]. No electrolyte paste was applied <strong>and</strong> the GSR electrodeswere directly fixed <strong>on</strong> the skin using medical tape. Proper c<strong>on</strong>tact with the skin was c<strong>on</strong>trolledusing the indicators implemented in the Actiview software <strong>and</strong> by verifying the existence of anEDR when the participant is stimulated by a surprising event (clap of the h<strong>and</strong>s hidden to theparticipant). Since the exosomatic setting implies that a small current is applied <strong>on</strong> the skin, theCMS / DRL electrodes have to be c<strong>on</strong>nected. If EEG was recorded simultaneously with GSR, thest<strong>and</strong>ard positi<strong>on</strong> was adopted (see secti<strong>on</strong> 3.2.1); otherwise the CMS / DRL electrodes werepositi<strong>on</strong>ed close to each other, <strong>on</strong> the hypothenar eminences of the same h<strong>and</strong> as the GSRelectrodes.2 GSRelectrodesPletysmographTemperatureRespirati<strong>on</strong> belttied around theabdomenCMS / DRL positi<strong>on</strong>Figure 3.3. Pictures <strong>and</strong> positi<strong>on</strong>s of the sensors used to m<strong>on</strong>itor peripheral activity. The CMS / DRL positi<strong>on</strong>was used <strong>on</strong>ly in the case where EEG activity was not m<strong>on</strong>itored simultaneously with peripheral activity.BVP (Blood Volume Pulse) was recorded by using a photoplethysmograph (later <strong>on</strong> calledplethysmograph) that emits an infrared light <strong>and</strong> record the amount of light reflected by the skin(see Secti<strong>on</strong> 2.2.2.b). The plethysmograph was often clipped <strong>on</strong> the participant’s thumb except ifit did not fit or if bad signal quality was obtained at that place. In this case the little finger wasused as the alternate positi<strong>on</strong>. The quality of the signal was verified by c<strong>on</strong>trolling the existenceof a pulse including the usual peaks observed in BVP signals (see Secti<strong>on</strong> 3.3.3).Respirati<strong>on</strong> was m<strong>on</strong>itored using a respirati<strong>on</strong> belt tied around the abdomen of participants.When allowed by the participant, the respirati<strong>on</strong> belt was placed directly in c<strong>on</strong>tact with the skinto avoid artifacts due to clothes movements. The respirati<strong>on</strong> belt is useful to determine therespirati<strong>on</strong> rate as well as deep breath depth. Since the chest expansi<strong>on</strong> was not m<strong>on</strong>itoredtogether with the abdomen expansi<strong>on</strong> it was not possible to determine the quantity of expired <strong>and</strong>inspired air (see Secti<strong>on</strong> 2.2.2.d). Between chest <strong>and</strong> abdomen positi<strong>on</strong>s, the abdomen waschosen because less noise was observed in the respirati<strong>on</strong> signals at that positi<strong>on</strong>. Be<str<strong>on</strong>g>for</str<strong>on</strong>g>e the54


Physiological signals recording <strong>and</strong> processingbeginning of each experiment, the quality of the respirati<strong>on</strong> signals was c<strong>on</strong>trolled by observingtheir shape in different respirati<strong>on</strong> c<strong>on</strong>diti<strong>on</strong>s (slow / normal / fast respirati<strong>on</strong>, deep breath).Finally a temperature sensor was attached <strong>on</strong> the distal phalange of the ring finger with medicaltape. Temperature measurements were assumed to be correct if they were around 33 °C.3.2 Signals acquisiti<strong>on</strong> <strong>and</strong> preprocessing3.2.1 Signal acquisiti<strong>on</strong>All the signals were acquired at a sampling rate of 1024 Hz. This high sampling rate was chosento have high time resoluti<strong>on</strong> EEG signals <strong>and</strong> it was applied to the other signal sources as wellbecause they were plugged <strong>on</strong> the same acquisiti<strong>on</strong> system. At that sampling rate the b<strong>and</strong>widthwas of 268Hz due to an antialiasing filter that is applied to the signals by the Actiview systembe<str<strong>on</strong>g>for</str<strong>on</strong>g>e storage <strong>on</strong> the hard drive. The complete b<strong>and</strong>with was not used in the present studies;nevertheless it allows per<str<strong>on</strong>g>for</str<strong>on</strong>g>ming high frequency EEG analysis <str<strong>on</strong>g>for</str<strong>on</strong>g> future studies.The physiological activity of each participant was recorded <str<strong>on</strong>g>for</str<strong>on</strong>g> the complete durati<strong>on</strong> of theprotocols. The obtained signals were then segmented in several trials (or epochs) each <strong>on</strong>e beingassociated to a given stimulus. The starting time of each stimulus could be retrieved thanks to thetriggers while the durati<strong>on</strong> of a trial was defined by the protocols. Be<str<strong>on</strong>g>for</str<strong>on</strong>g>e extracting physiologicalfeatures that are related to emoti<strong>on</strong>al activity, three preprocessing steps were applied to thesignals: EEG <strong>and</strong> peripheral signals were denoised, EEG signals were re-referenced <strong>and</strong> the heartrate (HR) signal was computed from the n<strong>on</strong>-filtered BVP signal.3.2.2 DenoisingAs a first step, the EEG signals were filtered by a 2-47 Hz Equiripple b<strong>and</strong> pass filter (Table 3.1).This filter was applied to remove the DC offset of each electrode, drifts due to the difference ofelectrode impedance over time <strong>and</strong> power lines 50 Hz noise. This b<strong>and</strong> pass filter also allowspreserving frequency b<strong>and</strong>s of interest <str<strong>on</strong>g>for</str<strong>on</strong>g> the study of emoti<strong>on</strong>al processes.The peripheral signals were filtered by a moving average filter to remove noise. For this purposewe used filters of length 512 samples <str<strong>on</strong>g>for</str<strong>on</strong>g> GSR <strong>and</strong> temperature, 128 <str<strong>on</strong>g>for</str<strong>on</strong>g> BVP, <strong>and</strong> 256 <str<strong>on</strong>g>for</str<strong>on</strong>g>respirati<strong>on</strong> (Table 3.1). Those different lengths were chosen to remove high frequencies withoutcorrupting oscillati<strong>on</strong>s of interest in the different signals.All the signals were filtered using the filtfilt functi<strong>on</strong> from the signal processing Matlab toolbox(v. 6.2.1) which processes the input signal in both the <str<strong>on</strong>g>for</str<strong>on</strong>g>ward <strong>and</strong> reverse directi<strong>on</strong>s. Thisfuncti<strong>on</strong> allows per<str<strong>on</strong>g>for</str<strong>on</strong>g>ming a zero-phase filtering.55


Chapter 3High pass(-3dB)Low pass(-3dB)EEG 2 Hz 47 HzGSR - 0.9 HzBVP - 3.5 HzRespirati<strong>on</strong> - 1.7 HzTemperature - 0.9 HzTable 3.1. Low <strong>and</strong> high pass cutoff frequencies at -3dB <str<strong>on</strong>g>for</str<strong>on</strong>g> the different filters.3.2.3 EEG re-referencingSince with the Biosemi system the original reference (signals are originally referenced to theCMS electrode) provides a poor signal to noise ratio, it is necessary to re-reference themafterward. To obtain a Laplacian reference the following Laplacian operator was applied to eachelectrode i:1x( n) x ( n) x ( n)(3.1)i i jNij Neigi ()where xiis the CMS referenced signal of electrode i, x i the Laplacian referenced signal, n thesample number, Neig(i) the neighbors electrodes of electrode i <strong>and</strong> N i the size of thisneighborhood. The neighborhood of an electrode was defined according to the Appendix B.563.2.4 HR computati<strong>on</strong>As stated in Secti<strong>on</strong> 2.2.2.c an HR signal can be inferred from the BVP signal recorded by apletysmograph. However, computing HR from a BVP signal is less reliable than from an ECGsince the vaso-c<strong>on</strong>structi<strong>on</strong> can influence the shape <strong>and</strong> timing of the heart pulses. As can be seenfrom Figure 3.4, the BVP signal is periodic because the blood pressure changes with heartc<strong>on</strong>tracti<strong>on</strong>s. Two points of interest can be identified in this signal: the foot of the systolicupstroke <strong>and</strong> the systolic peak both due to <strong>on</strong>e of the heart c<strong>on</strong>tracti<strong>on</strong>.A method to determine HR from a BVP signal is proposed in [128]. This method is based <strong>on</strong> acomplex analysis that identifies the systolic peaks as heart beats <strong>and</strong> requires recordings of l<strong>on</strong>gdurati<strong>on</strong>. For this study, this method was not used because:- as can be seen from Figure 3.4 (right) it is sometimes difficult to identify which peak iswhich in a pulse, especially when the blood pressure is str<strong>on</strong>gly increasing <strong>and</strong> decreasingor if the signal is of bad quality (noise due to movements, sensor badly placed or moved,etc.);- in our experiments the trials generally lasted less than 10 sec<strong>on</strong>ds, the method was thusnot adequate <str<strong>on</strong>g>for</str<strong>on</strong>g> these signals.


Physiological signals recording <strong>and</strong> processingSystolicpeakTidal peakFoot of thesystolicupstrokeTime (samples)Time (samples)Figure 3.4. The heart waves in a BVP signal. (Left) Three pulses of the BVP signal with the different peaks,(rigth) example of a pulse where it is difficult to identify the different peaks.In order to compute HR from signals of short durati<strong>on</strong>s without using the systolic peak, a methodbased <strong>on</strong> the detecti<strong>on</strong> of the foot of the systolic upstroke was implemented. The use of this point<str<strong>on</strong>g>for</str<strong>on</strong>g> identificati<strong>on</strong> of a heart beat is motivated by its frequent use <str<strong>on</strong>g>for</str<strong>on</strong>g> pulse wave velocitycomputati<strong>on</strong> [96] (i.e. the time elapsed between the heart beat <strong>and</strong> the corresp<strong>on</strong>ding wave in ablood pressure signal). The developed method is composed of the following steps:1. the linear trend of the BVP signal was removed from each trial to attenuate the effects ofstr<strong>on</strong>g increase <strong>and</strong> decrease of blood pressure;2. heart beats were assumed to be the local minima of the signal which were obtained byfinding samples were the derivative is zero <strong>and</strong> the amplitude is switching from adecrease to an increase;3. in the case where two such beats fall in the same interval of 0.5 sec<strong>on</strong>d then <strong>on</strong>ly the beatthat corresp<strong>on</strong>ds to the highest increasing BVP derivative is kept. The 0.5 sec<strong>on</strong>d intervalwas chosen based <strong>on</strong> the assumpti<strong>on</strong> that the HR will not exceed 120 beats per minutes(BPM) which is somehow reas<strong>on</strong>able since in all the protocols participants were sitting infr<strong>on</strong>t of a computer screen without per<str<strong>on</strong>g>for</str<strong>on</strong>g>ming any significant physical activity;4. the interbeat intervals (IBI) were computed as the time elapsed between two c<strong>on</strong>secutivebeats which could be c<strong>on</strong>verted to B-1 HR values corresp<strong>on</strong>ding to the B detected heartbeats. The time stamp of each HR value was placed in the middle of the corresp<strong>on</strong>dingIBI interval.57


Chapter 3Figure 3.5. Example of the beat detecti<strong>on</strong> <strong>and</strong> HR computati<strong>on</strong> algorithm <strong>on</strong> a 9 sec<strong>on</strong>ds signal. The HR signalis represented as a staircase functi<strong>on</strong> with the length of a step corresp<strong>on</strong>ding to the durati<strong>on</strong> of an IBI.As can be seen from Figure 3.5 this method per<str<strong>on</strong>g>for</str<strong>on</strong>g>med fairly well <strong>on</strong> short durati<strong>on</strong> trials (lessthan 10 sec<strong>on</strong>ds) but it was found to be less reliable <strong>on</strong> signals from <strong>on</strong>e of the protocols wherethe length of a trial was l<strong>on</strong>ger (5 minutes) <strong>and</strong> signals were noisier. To improve the reliability ofthe peak detecti<strong>on</strong>, an algorithm was designed to detect <strong>and</strong> correct the falsely detected heartbeats a posteriori. It is composed of two main steps described below.581. Detecti<strong>on</strong> <strong>and</strong> correcti<strong>on</strong> of false positive peaks (a beat is detected but this is not a true<strong>on</strong>e):a. <str<strong>on</strong>g>for</str<strong>on</strong>g> the ith IBI the median m i of the 5 precedent IBI’s was computed (except <str<strong>on</strong>g>for</str<strong>on</strong>g> the5 first IBI);b. if mIBIi () then the ith IBI is c<strong>on</strong>sidered as corrupted, being a parameteriof the algorithm;c. <str<strong>on</strong>g>for</str<strong>on</strong>g> each corrupted IBI check if removing <strong>on</strong>e of the two corresp<strong>on</strong>ding peak willsolve the problem, in this case this peak is removed.2. Detecti<strong>on</strong> <strong>and</strong> correcti<strong>on</strong> of false negative peaks (a beat has not been identified):a. similarly to step 1.b. if mIBIi () then the ith IBI is c<strong>on</strong>sidered ascorrupted;ib. <str<strong>on</strong>g>for</str<strong>on</strong>g> each corrupted IBI i the number P i of peaks to add was determined byP round( IBI( i)/ m ) 1; if adding those peaks does not violate the c<strong>on</strong>straintsii


Physiological signals recording <strong>and</strong> processing1.a <strong>and</strong> 2.a then the peaks were added in this IBI to c<strong>on</strong>struct P i +1 new IBI’s withequal durati<strong>on</strong>.The parameter was empirically set to 0.2 sec<strong>on</strong>ds since it corresp<strong>on</strong>ds to reas<strong>on</strong>able changes inHR <strong>and</strong> it detected nearly all the false positive <strong>and</strong> false negative peaks.3.3 Characterizati<strong>on</strong> of physiological activityFor each trial the physiological activity was characterized by computing several features from thesignals. It is then possible to c<strong>on</strong>catenate the different features to c<strong>on</strong>struct a feature vectorassociated to a trial. The current secti<strong>on</strong> explains <str<strong>on</strong>g>for</str<strong>on</strong>g> each feature why it has been chosen as wellas the time c<strong>on</strong>straints necessary <str<strong>on</strong>g>for</str<strong>on</strong>g> their accurate computati<strong>on</strong>.As explained in Secti<strong>on</strong> 2.2.3 it is important to use a baseline to account <str<strong>on</strong>g>for</str<strong>on</strong>g> inter <strong>and</strong> intraparticipant variability. However, most of the classifiers used in this thesis are participant specificwhich means that a classifier was trained <str<strong>on</strong>g>for</str<strong>on</strong>g> each participant. There is thus no need to account<str<strong>on</strong>g>for</str<strong>on</strong>g> inter-participant variability. Moreover, the physiological activity of each participant wasrecorded in a single sessi<strong>on</strong> within a c<strong>on</strong>trolled c<strong>on</strong>text (stable envir<strong>on</strong>ment temperature, nopossibility to eat, etc.) alleviating the need to account <str<strong>on</strong>g>for</str<strong>on</strong>g> intra-participant variability. Whilebaselines were computed <str<strong>on</strong>g>for</str<strong>on</strong>g> all protocols, subtracting baselines to the signals did not increase theaccuracy of participant specific classificati<strong>on</strong>. For this reas<strong>on</strong> the baselines were not used except<str<strong>on</strong>g>for</str<strong>on</strong>g> the video-game protocol presented in Chapter 7 because in this protocol a classifier wasdesigned independently of the participant. The baselines computed <str<strong>on</strong>g>for</str<strong>on</strong>g> this protocol are presentedin Chapter 7.3.3.1 St<strong>and</strong>ard featuresThis secti<strong>on</strong> described the physiological features that can be c<strong>on</strong>sidered as st<strong>and</strong>ard because oftheir frequent use <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment [87, 93, 106, 113, 123]. Only the features that wereused in this study are described.By assuming that each measured signal is generated by a Gaussian process with independent <strong>and</strong>identically distributed samples, the two physiological features that can be used to characterize aphysiological signal are its mean <strong>and</strong> its st<strong>and</strong>ard deviati<strong>on</strong>:N1 sx xn ( )(3.2)Ns n1x1NN s2 ( xn ( ) x)(3.3)s n159


Chapter 3where x(n) is the value of the signal x (an EEG electrode signal, GSR signal, etc.) at sample n <strong>and</strong>N s is the number of samples in a trial.In order to evaluate the trend of a signal x over a trial, the average of the signal derivative can becomputed as:N 11 s x( Ns) x(1)x ( xn ( 1) xn( )) (3.4)N 1 N 1sn1Finally the maximum <strong>and</strong> minimum of a signal can also provide in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> c<strong>on</strong>cerning therange of the signal amplitude:MinMaxxxs min xn ( )(3.5)n max xn ( )(3.6)Those features are quite general <strong>and</strong> can be applied to a wide range of physiological signals(EEG, GSR, EMG, etc.). Their usefulness (or weakness) will be discussed in the Secti<strong>on</strong> 3.4.2 <str<strong>on</strong>g>for</str<strong>on</strong>g>the different types of signal.3.3.2 Advanced featuresThis secti<strong>on</strong> proposes specific features <str<strong>on</strong>g>for</str<strong>on</strong>g> each signal type based <strong>on</strong> the variables that are oftenanalyzed in psycho-physiology.a. EEGThe cognitive theory of emoti<strong>on</strong>s provides a str<strong>on</strong>g motivati<strong>on</strong> to go toward emoti<strong>on</strong> assessmentusing signals from the CNS. As described in Secti<strong>on</strong> 2.2.1.c, several researchers have shown theinvolvement of brain structures in emoti<strong>on</strong>al processes. When using EEG to record emoti<strong>on</strong>alactivity, most of the results were obtained by comparing the energy in different frequency b<strong>and</strong>s.For instance, Davids<strong>on</strong> [81] dem<strong>on</strong>strated the lateralizati<strong>on</strong> of the fr<strong>on</strong>tal cortex by studying theenergy of the EEG alpha waves. Aftanas et al. [82] reported energy differences between more orless arousing visual stimuli (images from the IAPS). Those last differences were observed fromthe energy in theta b<strong>and</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> the parietal <strong>and</strong> occipital areas, alpha b<strong>and</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> the fr<strong>on</strong>tal areas <strong>and</strong>gamma b<strong>and</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> the complete scalp. However, energy is not the <strong>on</strong>ly feature that can be used <str<strong>on</strong>g>for</str<strong>on</strong>g>the purpose of emoti<strong>on</strong> assessment since studies also dem<strong>on</strong>strated that there are specific patternsof synchr<strong>on</strong>izati<strong>on</strong> between brain areas during emoti<strong>on</strong>al processes [61, 129]. For those reas<strong>on</strong>sdifferent types of features were defined to characterize EEG signals. They were regrouped in fivefeature sets according to the type of features extracted.n60


Physiological signals recording <strong>and</strong> processingEnergyThis set of features, named EEG_FFT, was defined to represent the energy of EEG signals infrequency b<strong>and</strong>s known to be related to emoti<strong>on</strong>al processes [81, 82]. For each electrode i, theenergy in the different frequency b<strong>and</strong>s displayed in Table 3.2 was computed over all samplesbel<strong>on</strong>ging to a specific trial, using the Fast Fourier Trans<str<strong>on</strong>g>for</str<strong>on</strong>g>m (FFT) algorithm. This was d<strong>on</strong>eunder the assumpti<strong>on</strong> that the EEG signals are stati<strong>on</strong>ary over the whole durati<strong>on</strong> of a trial.Feature <str<strong>on</strong>g>for</str<strong>on</strong>g>electrodes iFrequency b<strong>and</strong> i4-8 Hz i8-12 Hz i12-30 HzTable 3.2. The energy features computed <str<strong>on</strong>g>for</str<strong>on</strong>g> each electrode <strong>and</strong> the associated frequency b<strong>and</strong>s.Moreover, the following EEG_W feature was added because it is known to be related to cognitiveprocesses like workload, engagement, attenti<strong>on</strong> <strong>and</strong> fatigue [130-133]. To compute this featurethe energy in each frequency was summed over the N e electrodes:EEGii1_ W log( )Nei1Nei i(3.7)The feature vector f EEG_FFT , which corresp<strong>on</strong>ds to <strong>on</strong>e trial, is thus composed of 3.N e +1 features<strong>and</strong> the EEG_FFT feature set c<strong>on</strong>tains all the f EEG_FFT vectors.Lateralizati<strong>on</strong> of EEG alpha wavesThe results obtained in [19] are at the origin of the creati<strong>on</strong> of this feature set calledEEG_Lateral. This study dem<strong>on</strong>strated the correlati<strong>on</strong> between a EEG asymmetry score in thefr<strong>on</strong>tal regi<strong>on</strong> <strong>and</strong> reported measures of general tendency to approach or withdraw. Similarresults have also been reported, with less reproducibility, <str<strong>on</strong>g>for</str<strong>on</strong>g> stimuli of negative <strong>and</strong> positivevalence [58, 78]. The asymmetry score was computed according to the following <str<strong>on</strong>g>for</str<strong>on</strong>g>mula:AS log( P ) log( P )(3.8)Rwhere P R <strong>and</strong> P L are the power of the EEG signal in the alpha frequency b<strong>and</strong> (defined as the 8-13 Hz b<strong>and</strong>) respectively <str<strong>on</strong>g>for</str<strong>on</strong>g> a given right electrode <strong>and</strong> the left symmetrical electrode (<str<strong>on</strong>g>for</str<strong>on</strong>g>instance electrodes F4 <strong>and</strong> F3).L61


Chapter 3In this study the asymmetry score was computed <str<strong>on</strong>g>for</str<strong>on</strong>g> several pairs of electrodes, including n<strong>on</strong>fr<strong>on</strong>talelectrodes.Asymmetry score Right electrode Left electrodeAS_1 Fp2 Fp1AS_2 AF4 AF3AS_3 F4 F3AS_4 FC4 FC3AS_5 C4 C3AS_6 CP4 CP3AS_7 P4 P3AS_8 PO4 PO3AS_9 O2 O1Table 3.3. The pairs of electrodes used to compute 9 asymmetry scores.For each trial, the alpha power of a given electrode was computed from the FFT applied <strong>on</strong> thecomplete durati<strong>on</strong> of the trial; the asymmetry scores were then computed <strong>and</strong> a feature vectorf EEG_Lateral was c<strong>on</strong>structed by c<strong>on</strong>catenati<strong>on</strong> of the asymmetry scores:EEG _ Lateralf [ AS _1, , AS _ 9](3.9)Finally, the EEG_Lateral feature set is composed of the ensemble of the f EEG_Lateral feature vectorsobtained <str<strong>on</strong>g>for</str<strong>on</strong>g> each trial.Average energy over brain areasThis feature set was computed to assess emoti<strong>on</strong>s elicited by the visualizati<strong>on</strong> of images withemoti<strong>on</strong>al c<strong>on</strong>tent, as described in Chapter 5. The choice of those features was based <strong>on</strong> the studyof Aftanas et al. [82] who showed a correlati<strong>on</strong> between arousal elicited by images <strong>and</strong> resp<strong>on</strong>ses inparticular frequency b<strong>and</strong>s <strong>and</strong> brain areas. As can be seen in Figure 3.6, areas were defined asgroups of several electrodes (e.g. there are 6 electrodes in area PT, P, O). This figure also indicatethe name of the rhythms of interest (1, 2, , etc.) together with the associated frequency b<strong>and</strong>s.Power values of the 6 frequency b<strong>and</strong>s listed in Figure 3.6 were computed <str<strong>on</strong>g>for</str<strong>on</strong>g> each electrode <strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g>the whole durati<strong>on</strong> of a given trial using the FFT algorithm. As several electrodes are located in thesame area, the power over all these electrodes were averaged yielding a total of 6 features <str<strong>on</strong>g>for</str<strong>on</strong>g> thisEEG feature set (e.g. feature <strong>on</strong>e is the average power in b<strong>and</strong> 1 over all electrodes in areas PT, P,O). Most of the features c<strong>on</strong>cern the Occipital (O) lobe, which is not surprising since this lobecorresp<strong>on</strong>ds to the visual cortex <strong>and</strong> subjects were stimulated with pictures [82].62


Physiological signals recording <strong>and</strong> processingCPPTATFFATPTCPEEGfeatureLocati<strong>on</strong>areaFrequencyb<strong>and</strong>EEG_Area_1 [PT;P;O] 1 (4-6Hz)EEG_Area_2 [PT;P;O] 2 (6-8Hz)EEG_Area_3 [PT;P;O] (30-45Hz)EEG_Area_4 [AT;F] 2 (10-12Hz)EEG_Area_5 [AT;F;C] 1 (12-18Hz)EEG_Area_6 [C;PT;P;O] 3 (22-30Hz)OOFigure 3.6. Top head view with EEG electrode locati<strong>on</strong>s <strong>and</strong> corresp<strong>on</strong>ding frequency b<strong>and</strong>s (from [82]).For each trial a feature vector f EEG_Area was c<strong>on</strong>structed by c<strong>on</strong>catenating the 6 EEG features:EEG _ Areaf EEG Area EEG Area(3.10)[ _ _1, , _ _ 6]<strong>and</strong> the EEG_Area feature set was defined as the set of all vectors computed from the different trialsof a given protocol.SpectrogramIn the two feature sets, energy features were computed from EEG signals by applying the FFTalgorithm <strong>on</strong> the whole durati<strong>on</strong> of a trial. This could be d<strong>on</strong>e reliably under the assumpti<strong>on</strong> thatthe EEG signals are stati<strong>on</strong>ary <str<strong>on</strong>g>for</str<strong>on</strong>g> the durati<strong>on</strong> of each trial. However this is rarely true sinceEEG signals can <strong>on</strong>ly be c<strong>on</strong>sidered as stati<strong>on</strong>ary <strong>on</strong> short periods of time. This comment alsoapplies <str<strong>on</strong>g>for</str<strong>on</strong>g> the computati<strong>on</strong> of mean <strong>and</strong> variance features. To solve this issue the followingfeature set was defined under the assumpti<strong>on</strong> that EEG signals are stati<strong>on</strong>ary <strong>on</strong> short timewindows of 0.5 sec<strong>on</strong>ds.This set of EEG features was extracted by computing the Short-Time Fourier Trans<str<strong>on</strong>g>for</str<strong>on</strong>g>m (STFT)<str<strong>on</strong>g>for</str<strong>on</strong>g> each electrode with a sliding window of 512 samples <strong>and</strong> 50% overlap between c<strong>on</strong>secutivewindows. This window size was chosen in order to have a frequency resoluti<strong>on</strong> of f = 2Hzwhich allows to separate the different rhythms observed in EEG signals while maintaining a timeresoluti<strong>on</strong> of 0.5 sec<strong>on</strong>d. For each of the spectrograms (<strong>on</strong>e per electrode), we selected 9frequency b<strong>and</strong>s of 2Hz ranging from 4Hz to 22Hz. This range was chosen because it includesmost of the frequency b<strong>and</strong>s used <str<strong>on</strong>g>for</str<strong>on</strong>g> the computati<strong>on</strong> of the EEG_FFT <strong>and</strong> EEG_Area featuresset.The f EEG_STFT feature vector <str<strong>on</strong>g>for</str<strong>on</strong>g> a given trial is then c<strong>on</strong>structed by c<strong>on</strong>catenating all the powervalues of the 9 frequency b<strong>and</strong>s at the different time frames <strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> each electrode. The length off EEG_STFT is thus:63


Chapter 3length N TEEG _ STFT( f ) 9 e4 1where 9 is the number of frequency b<strong>and</strong>s, N e the number of electrodes <strong>and</strong> 4T 1 gives thenumber of time frames <str<strong>on</strong>g>for</str<strong>on</strong>g> a trial of durati<strong>on</strong> T sec<strong>on</strong>ds. The feature set composed of all thef EEG_STFT feature vectors (<strong>on</strong>e per trial) is named EEG_STFT.Mutual In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> featuresIn this feature set, mutual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> (MI) between pairs of electrodes is proposed as a measureof statistical dependencies between different areas of the brain. This set of features was motivatedby studies that dem<strong>on</strong>strated synchr<strong>on</strong>izati<strong>on</strong> of brains areas in emoti<strong>on</strong>al processes [61, 129].With the assumpti<strong>on</strong> that the signal x i of electrode i <str<strong>on</strong>g>for</str<strong>on</strong>g> a given trial is a stochastic process withprobability mass functi<strong>on</strong> P(X i ), the mutual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> between electrodes i <strong>and</strong> j <str<strong>on</strong>g>for</str<strong>on</strong>g> this trial isexpressed as:I( X ; X ) H( X ) H( X | X )i j i i j (3.11)H( X ) P( X x)log( P( X x))i i i i ixiH( X / X ) P( X x, X x )log( P( X x / X x ))i j i i j j i i j jx , xijwhere H(X i ) <strong>and</strong> H(X i | X j ) are respectively the entropy <strong>and</strong> c<strong>on</strong>diti<strong>on</strong>al entropy of r<strong>and</strong>omvariables X i <strong>and</strong> X j . Mutual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> was computed using Moddemeijer’s Matlab toolbox[134] 8 that estimates the different distributi<strong>on</strong>s based <strong>on</strong> histograms <strong>and</strong> automatically determinesan appropriate bin size.The MI feature vector f EEG_MI of this trial is then c<strong>on</strong>structed by c<strong>on</strong>catenati<strong>on</strong> of mutualin<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> between each pairs of electrodes:[ ( , ) ( , ), ( , ) ( , ), ( , )]EEG _ MIf I X1 X2 I X1 XN I Xi Xi1 I Xi XN I XN 1XN(3.12)e e e eThe total number of features of a trial <str<strong>on</strong>g>for</str<strong>on</strong>g> N e electrodes is then:set c<strong>on</strong>taining all feature vectors was named EEG_MI.N e1i1( Ne1)Ni 2e. The feature8 available at http://www.cs.rug.nl/~rudy/matlab/ (retrieved <strong>on</strong> 27 April 2009)64


Physiological signals recording <strong>and</strong> processingb. Galvanic Skin Resp<strong>on</strong>se (GSR)The mean skin resistance over a trial has been shown to be correlated with the arousal of astimulus [7] <strong>and</strong> has been widely used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment from peripheral physiologicalsignals. An aroused emoti<strong>on</strong> should also induce a decrease in the GSR signal. For those reas<strong>on</strong>smean skin resistance <strong>and</strong> the trend of the GSR signal were frequently added to the peripheralfeature sets. However, as described in Secti<strong>on</strong> 2.2.2.a, the speed of the fall of a GSR signal is alsoof importance to characterize the EDA. To approximate this value the average decrease rateduring decay time was defined as follows:fDecRateGSR1 GSR'( n)(3.13)Nn n/ GSR'( n) 0with N n being the number of samples where the GSR derivative GSR ' is negative. The proporti<strong>on</strong>of negative samples in the derivative was also computed to characterize the durati<strong>on</strong> of GSRfalls:fDecTimeGSRNnN 1s(3.14)While these features can correctly characterize GSR signals when it is expected that <strong>on</strong>ly <strong>on</strong>eEDR occurs, it is important to count the number of EDR’s if several resp<strong>on</strong>ses occur in a singletrial. For this purpose an automatic EDR detecti<strong>on</strong> algorithm called nbPeaks was designed. In afirst step this algorithm identifies c<strong>on</strong>secutive high <strong>and</strong> low peaks of a GSR signal by finding thesign changes of the signal derivative. High peaks were identified as potentially being thebeginning of an EDR whereas the low peaks were identified as the potential apex of a resp<strong>on</strong>se.Finally a resp<strong>on</strong>se was counted if the following criteria were met:- the difference in amplitude between the beginning <strong>and</strong> the apex of a resp<strong>on</strong>se is higherthan 200 Ohms;- the time elapsed between the beginning <strong>and</strong> the apex of a resp<strong>on</strong>se is between 1 <strong>and</strong> 5sec<strong>on</strong>ds.Thus the last feature computed <str<strong>on</strong>g>for</str<strong>on</strong>g> the GSR signal is:NbPeaksfGSR nbPeaks( GSR)(3.15)C<strong>on</strong>cerning the time aspects, the EDR is known to occur from 1 to 4 sec<strong>on</strong>ds after a stimulus[88]. It is thus important to record a GSR signal <str<strong>on</strong>g>for</str<strong>on</strong>g> more than 4 sec<strong>on</strong>ds after the presentati<strong>on</strong> ofan emoti<strong>on</strong>al stimulus to be sure that at least part of the reacti<strong>on</strong> is recorded by the sensor.65


Chapter 3c. Blood volume pulse (BVP)The BVP signal given by the pletysmograph is a relative measure of the blood pressure (BP): it iscorrelated with the BP but does not measure its exact value. Since variati<strong>on</strong>s of BP <strong>and</strong> bloodflow are known to be associated to several emoti<strong>on</strong>s (Sinha, Healey, Secti<strong>on</strong> 2.2.2b) the featurescomputed from the BVP signal were the mean, the st<strong>and</strong>ard deviati<strong>on</strong> <strong>and</strong> minimum / maximumvalues as described in Secti<strong>on</strong> 3.4.1.d. Heart rate (HR)The increase <strong>and</strong> decrease of HR is associated to many emoti<strong>on</strong>s (Secti<strong>on</strong> 2.2.2.c) thus theaverage HR computed over the whole durati<strong>on</strong> of an emoti<strong>on</strong>al stimulus is a feature that can beused <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment. Another variable of interest <str<strong>on</strong>g>for</str<strong>on</strong>g> the study of emoti<strong>on</strong>s is the HeartRate Variability (HRV). As detailed in Secti<strong>on</strong> 2.2.2.c the HRV can be determined from the HRpower spectrum <strong>and</strong> by computing HR st<strong>and</strong>ard deviati<strong>on</strong>.Three comp<strong>on</strong>ents in the HR spectrum (frequency b<strong>and</strong>s) of interest are generally c<strong>on</strong>sidered tocharacterize HRV: a High Frequency b<strong>and</strong> (HF, 0.15Hz-1Hz), a Low Frequency b<strong>and</strong> (LF,0.05Hz-0.15Hz) <strong>and</strong> a Very Low Frequency b<strong>and</strong> (VLF, 0.0033Hz-0.05Hz). As a c<strong>on</strong>sequence, itis necessary to have signals of significant durati<strong>on</strong> to correctly record HRV comp<strong>on</strong>ents. Forinstance the committee report of the Society <str<strong>on</strong>g>for</str<strong>on</strong>g> Psychophysiological Research [96] recommendsusing epochs of at least 1 minute to reliably compute the HF comp<strong>on</strong>ent <strong>and</strong> 2 minutes <str<strong>on</strong>g>for</str<strong>on</strong>g> the LFcomp<strong>on</strong>ent. Those durati<strong>on</strong>s corresp<strong>on</strong>d to approximately 6-10 times the length of the wave withthe lowest frequency (0.15Hz <str<strong>on</strong>g>for</str<strong>on</strong>g> HF <strong>and</strong> 0.05Hz <str<strong>on</strong>g>for</str<strong>on</strong>g> LF). Actually, it is possible to determine(less reliably) the energy in the HF b<strong>and</strong> <strong>on</strong> epochs of 6.6 sec<strong>on</strong>ds <strong>and</strong> in the LF b<strong>and</strong> <strong>on</strong> epochsof 20 sec<strong>on</strong>ds. If HRV is estimated using the st<strong>and</strong>ard deviati<strong>on</strong> of heart rate, then it iscomputable <strong>on</strong> any period of time including at least 2 HR values (3 heart beats, which generallycorresp<strong>on</strong>ds to less than 3 sec<strong>on</strong>ds); the result will however <strong>on</strong>ly provide in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> about theenergy in frequency b<strong>and</strong>s limited by the durati<strong>on</strong> of the epoch.The VLF comp<strong>on</strong>ent was not investigated in this study since it needs too l<strong>on</strong>g trial durati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> itscomputati<strong>on</strong>. Be<str<strong>on</strong>g>for</str<strong>on</strong>g>e computing energy in the HF <strong>and</strong> LF frequency b<strong>and</strong>s the HR signal wasinterpolated with a cubic interpolati<strong>on</strong> to obtain a new signal sampled at 1024Hz. This first stepallows to:- trans<str<strong>on</strong>g>for</str<strong>on</strong>g>m the original signal that has a low sampling rate <strong>and</strong> which is not sampled atregular intervals into a regularly sampled signal with a high sampling rate;- obtain a signal with the same sampling rate than the others which facilitates thecomparis<strong>on</strong> of the signals over time (correlati<strong>on</strong> measures, etc.).66


Physiological signals recording <strong>and</strong> processingThe FFT algorithm was then applied to the interpolated HR signal to compute the power in thetwo frequency b<strong>and</strong>s of interest. Finally, the ratio between the power in the HF <strong>and</strong> LF frequencyb<strong>and</strong>s was computed to reflect the cardiac balance between the sympathetic <strong>and</strong> parasympatheticactivity (see Secti<strong>on</strong> 2.2.2.c). Table 3.2 summarizes the three features computed from the HRpower spectrum.e. Respirati<strong>on</strong>Feature name Descripti<strong>on</strong> / <str<strong>on</strong>g>for</str<strong>on</strong>g>mulaLFfHRPower of HR in [0.05-0.15Hz]HFfHRPower of HR in [0.15-1Hz]LF / HFfHRffLFHRHFHRTable 3.4. The three features computed from the HRThe respirati<strong>on</strong> rate ranges from 0.1 to 0.35 breaths / sec<strong>on</strong>d at rest, while it can reach 0.7 breaths/ sec<strong>on</strong>d during exercise. Moreover, if respirati<strong>on</strong> is measured with a belt, irregular respirati<strong>on</strong>lead to energy increases in higher frequency b<strong>and</strong>s. Laughing also affects the respirati<strong>on</strong> patternby introducing high-frequency fluctuati<strong>on</strong>s <strong>on</strong> the recorded signal. To measure the mainrespirati<strong>on</strong> frequency (i.e. respirati<strong>on</strong> rate) at rest it is thus necessary to have a signal of at least10 sec<strong>on</strong>ds (based <strong>on</strong> the lower bound of the respirati<strong>on</strong> rate). Nevertheless, shorter periods oftime could still c<strong>on</strong>tain interesting in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> since respirati<strong>on</strong> rate at rest can increase up to0.35 breaths / sec<strong>on</strong>d which can be measured <strong>on</strong> an epoch of around 3 sec<strong>on</strong>ds. To capture thosefluctuati<strong>on</strong>s, features from both the frequency <strong>and</strong> time domain are there<str<strong>on</strong>g>for</str<strong>on</strong>g>e used.Features of the frequency domain are obtained by computing the FFT of the original signal. Thensix features were computed to measure the power of the respirati<strong>on</strong> signal in c<strong>on</strong>secutivefrequency b<strong>and</strong>s. Table 3.5 summarizes the different frequency b<strong>and</strong>s of interest. Thosefrequency b<strong>and</strong>s were c<strong>on</strong>catenated in a feature vector calledPowfResp.Power feature Frequency b<strong>and</strong> Power feature Frequency b<strong>and</strong>f [0-0.25Hz]f [0.75-1Hz]Pow1RespPow2fResp[0.25-0.5Hz]Pow3fResp[0.5-0.75Hz]Pow4RespPow5fResp[0.75-1.25Hz]Pow6fResp[1.25-1.5Hz]Table 3.5. The power features computed from the respirati<strong>on</strong> signals <strong>and</strong> being part of thevector.PowfRespfeatureThe main frequency of the respirati<strong>on</strong> signal was c<strong>on</strong>sidered to represent the respirati<strong>on</strong> rate <strong>and</strong>is computed as:67


Chapter 3Ratef arg max P ( f )(3.16)Resp0.1f0.7Respwhere P Resp (f) is the power of the respirati<strong>on</strong> signal at frequency f. The precisi<strong>on</strong> of this featuredepends <strong>on</strong> the frequency resoluti<strong>on</strong> obtained from the FFT which depends <strong>on</strong> the durati<strong>on</strong> of thesignal used <str<strong>on</strong>g>for</str<strong>on</strong>g> the FFT computati<strong>on</strong>.Finally, to identify the deepest inhalati<strong>on</strong> or exhalati<strong>on</strong> in a respirati<strong>on</strong> signal Resp, the dynamicrange feature was computed as:f. TemperatureDRf max Respn ( ) min Respn ( )(3.17)RespnGenerally, changes in skin temperature are quite slow <strong>and</strong> difficult to detect <strong>on</strong> short timeperiods. However, since skin temperature is influenced by vasoc<strong>on</strong>stricti<strong>on</strong> <strong>and</strong> sweat, it ispossible to observe short thermal reacti<strong>on</strong>s of a few sec<strong>on</strong>ds. For instance, Ekman [5] showedthat the average value of skin temperature measured <strong>on</strong> a h<strong>and</strong> over an epoch of 10 sec<strong>on</strong>ds couldbe used to distinguish anger from fear <strong>and</strong> sadness. For this reas<strong>on</strong> mean temperature was used asa feature in this study. The average derivative of the temperature signal was also used to indicatethe trend of the signal.3.4 Ethical <strong>and</strong> privacy aspectsEthics is related to moral, law, h<strong>on</strong>esty <strong>and</strong> privacy issues [135]. It is not often that researchers incomputer science have to deal with those issues. However, there is an ever exp<strong>and</strong>ing trend inrecording <strong>and</strong> using pers<strong>on</strong>al data <str<strong>on</strong>g>for</str<strong>on</strong>g> various domains of applicati<strong>on</strong> such as sport, gaming <strong>and</strong>health. Due to its very private nature, this type of pers<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> requires that all necessarysteps be taken to ensure privacy. In order to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m experiments <strong>on</strong> physiological emoti<strong>on</strong>assessment it is necessary to acquire physiological data from pers<strong>on</strong>s while they are experiencingemoti<strong>on</strong>s. This raises several ethical issues since the human being is directly involved in thoseexperiments <strong>and</strong> sensitive data is acquired.This secti<strong>on</strong> describes all the steps that were taken to ensure ethical aspects. The bindingregulati<strong>on</strong>s are European (The European Charter of Fundamental Rights, Art. 3, 8, 13), Swiss(Swiss Federal law from June 19, 1992 <strong>on</strong> data protecti<strong>on</strong> (LPD)), as well as at the Universitylevel (University of Geneva regulati<strong>on</strong>s c<strong>on</strong>cerning integrity in research). Rules <strong>and</strong> key points toaddress were also derived from the European commissi<strong>on</strong> document c<strong>on</strong>cerning ethics <str<strong>on</strong>g>for</str<strong>on</strong>g>researchers [135].n68


Physiological signals recording <strong>and</strong> processingThe recordings were per<str<strong>on</strong>g>for</str<strong>on</strong>g>med with a limited number of volunteers (ranging from 4 to 11pers<strong>on</strong>s depending <strong>on</strong> the experiment). All of them were adults, researchers from our departmentor students interested in our work. It was c<strong>on</strong>trolled that participants were able to underst<strong>and</strong> <strong>and</strong>questi<strong>on</strong> the experiment as well as give their participati<strong>on</strong> approval <strong>on</strong> their own. The participantsalso needed to accept spending some time to do the experiment but no other selecti<strong>on</strong> criteri<strong>on</strong>was applied.All volunteers were carefully in<str<strong>on</strong>g>for</str<strong>on</strong>g>med about the experiment <strong>and</strong> had to sign a c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m(Appendix A) to ensure that they were aware of:- the purpose <strong>and</strong> scientific goals of the research (in human-computer interacti<strong>on</strong>);- how the experiment is per<str<strong>on</strong>g>for</str<strong>on</strong>g>med (sensors used, type of data recorded, descripti<strong>on</strong> of theprotocol, durati<strong>on</strong> of the experiment, possible income <str<strong>on</strong>g>for</str<strong>on</strong>g> the participant, etc.);- the possible risks incurred (see Appendix A <strong>and</strong> next paragraph);- the privacy issue, data being c<strong>on</strong>fidential <strong>and</strong> an<strong>on</strong>ymous;- the pers<strong>on</strong>s to c<strong>on</strong>tact in case of questi<strong>on</strong>s or problems (project leader <strong>and</strong> experimenters);- the possibility to stop the experiment at any time, <strong>and</strong> to have data deleted <strong>on</strong> request,both without any loss of benefits.Neither questi<strong>on</strong>s c<strong>on</strong>sidered as too pers<strong>on</strong>al nor medical questi<strong>on</strong>s were asked. The <strong>on</strong>lyincentives to participate in the studies were self-motivati<strong>on</strong> <strong>and</strong> curiosity. In <strong>on</strong>ly <strong>on</strong>e study a(small) financial reward was given to some participants according to their per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance (toincrease their motivati<strong>on</strong> see Chapter 7). Two copies of the (paper <strong>on</strong>ly) c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m weresigned: <strong>on</strong>e <str<strong>on</strong>g>for</str<strong>on</strong>g> the pers<strong>on</strong> participating to the experiment, <strong>on</strong>e <str<strong>on</strong>g>for</str<strong>on</strong>g> the project leader. Moreoverthe c<strong>on</strong>tent of this c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m was inspired from several sources like the binding regulati<strong>on</strong>sreferred above <strong>and</strong> other c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>ms used in medical <strong>and</strong> biological research.Regarding safety, all recordings of physiological signals were n<strong>on</strong>-invasive. The system usedin the studies (see Secti<strong>on</strong> 3.2) records physiological signals using active electrodes, which meanthat a small quantity of current is applied <strong>on</strong> the surface of the skin. To our knowledge thiscurrent is not harmful to the human. No counter-indicati<strong>on</strong>s are given by the manufacturer <strong>and</strong>supplier of the acquisiti<strong>on</strong> system. The electrodes <strong>and</strong> the A/D c<strong>on</strong>verter are galvanicaly isolated(using an optical fiber) from the rest of the acquisiti<strong>on</strong> system (PC), avoiding any risk ofelectrocuti<strong>on</strong> from the mains supply. These systems are routinely <strong>and</strong> safely used in manylaboratories worldwide, <str<strong>on</strong>g>for</str<strong>on</strong>g> a number of applicati<strong>on</strong>s including human-computer interacti<strong>on</strong>.Other threats could be headache due to a too l<strong>on</strong>g period wearing the fairly tight EEG caps <strong>and</strong>69


Chapter 3epileptic reacti<strong>on</strong>s due to the manipulati<strong>on</strong> of human computer interfaces (especially games). It isimportant to state that n<strong>on</strong>e of the experimenters have medical knowledge that could permit theincidental finding of diseases or abnormalities in the data; this was explicitly stated in the c<strong>on</strong>sent<str<strong>on</strong>g>for</str<strong>on</strong>g>m.Regarding privacy, all recordings were an<strong>on</strong>ymized by assigning a numerical code to eachuser, <strong>and</strong> stored accordingly (e.g. Participant 1, Participant 2, etc.). The names of the participantswere not stored in electr<strong>on</strong>ic <str<strong>on</strong>g>for</str<strong>on</strong>g>m. Only the project leader <strong>and</strong> the experimenter had access to theparticipant's identity. The relevant recorded data, according to the purpose of the experiment, arekept <str<strong>on</strong>g>for</str<strong>on</strong>g> processing. Only authorized researchers have access to the recorded an<strong>on</strong>ymous data (<str<strong>on</strong>g>for</str<strong>on</strong>g>electr<strong>on</strong>ically stored data: password <strong>and</strong> IP address restricti<strong>on</strong>). Data <strong>on</strong>ce recorded will not bemodified. The data is processed <strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g> the research purposes as is explained in the c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m<strong>and</strong> will not be used <str<strong>on</strong>g>for</str<strong>on</strong>g> any other research. All data can be erased <strong>on</strong> request from theparticipant. Further, it is not possible to determine the identity of a participant from thephysiological recordings we employ. The data is not shared with outside laboratories, <strong>and</strong> notused <str<strong>on</strong>g>for</str<strong>on</strong>g> any commercial purpose.Finally, this secti<strong>on</strong> does not aim at answering ethical questi<strong>on</strong>s c<strong>on</strong>cerning the potentialapplicati<strong>on</strong>s of this thesis. However, we believe that <str<strong>on</strong>g>for</str<strong>on</strong>g> each specific applicati<strong>on</strong> the benefit /burden balance should be c<strong>on</strong>sidered carefully.70


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentChapter 4Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment4.1 Classificati<strong>on</strong>Secti<strong>on</strong> 3.3 describes the different features extracted from the physiological signals to reflectemoti<strong>on</strong>al activity. The next step to go toward emoti<strong>on</strong> assessment c<strong>on</strong>sists in finding acomputati<strong>on</strong>al model that is able to associate a given instance of those features to an emoti<strong>on</strong>alstate.To obtain such a model it is possible to learn it from previously acquired data. In this thesisseveral protocols (described in the next chapters) were designed to acquire emoti<strong>on</strong>al data elicitedby different types of stimuli (images, recall of past emoti<strong>on</strong>al episodes, etc.). As explained inChapter 3, the signals acquired from those protocols were segmented in N trials, each <strong>on</strong>ecorresp<strong>on</strong>ding to an emoti<strong>on</strong>al state y i , <strong>and</strong> features were extracted <str<strong>on</strong>g>for</str<strong>on</strong>g> each trial. The result is adatabase <str<strong>on</strong>g>for</str<strong>on</strong>g> each protocol including:- a feature set F composed of N feature vectors f i , <strong>on</strong>e <str<strong>on</strong>g>for</str<strong>on</strong>g> each trial i, c<strong>on</strong>taining some ofthe features described in Secti<strong>on</strong> 3.3 (a feature vector f i can then be regarded as aninstance or sample of the feature set F);- a vector y with each value y i being the emoti<strong>on</strong>al state associated to trial (or sample) i.On such a database, supervised learning algorithms [119, 120] can be applied to obtain acomputati<strong>on</strong>al model that transduces the values of feature vector f i into an estimated emoti<strong>on</strong>alstate y ˆi. In supervised learning such a model is learned from a database that c<strong>on</strong>tains both afeature set F <strong>and</strong> the associated labels vector y. However, in order to c<strong>on</strong>struct the y vector it isfirst necessary to define what an emoti<strong>on</strong>al state is <strong>and</strong> how to determine the real emoti<strong>on</strong>associated to a trial. This is known as the ground-truth c<strong>on</strong>structi<strong>on</strong> <strong>and</strong> is detailed in Secti<strong>on</strong>4.1.1. Secti<strong>on</strong> 4.1.2 explains the methods employed to validate the models learned from the data<strong>and</strong> Secti<strong>on</strong> 4.1.3 presents the different supervised algorithms used to learn the model.4.1.1 Ground-truth definiti<strong>on</strong>Defining a ground-truth <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of emoti<strong>on</strong> assessment str<strong>on</strong>gly depends <strong>on</strong> the protocolused to record emoti<strong>on</strong>al reacti<strong>on</strong>s. There are actually two ways to elicit emoti<strong>on</strong>s (Secti<strong>on</strong> 2.3):- by asking the participant to self-generate a given emoti<strong>on</strong>;- by stimulating him / her with material c<strong>on</strong>taining emoti<strong>on</strong>al c<strong>on</strong>tent (images, sounds,video clips, etc.).71


Chapter 4In the first case the annotati<strong>on</strong> is straight<str<strong>on</strong>g>for</str<strong>on</strong>g>ward since a trial could be annotated with the emoti<strong>on</strong>that the participant was supposed to express. However, assumes that the requested emoti<strong>on</strong>s weresuccessfully elicited, <strong>and</strong> thus requires appropriate c<strong>on</strong>trol of the elicitati<strong>on</strong> procedure.In the sec<strong>on</strong>d case, the annotati<strong>on</strong> of each trial (corresp<strong>on</strong>ding to a stimulus) can be d<strong>on</strong>e byeither <strong>on</strong>e of the following three methods:- by a-priori defining the emoti<strong>on</strong>al label of each stimulus;- by determining the elicited emoti<strong>on</strong> from the observati<strong>on</strong> of the participant’s emoti<strong>on</strong>alexpressi<strong>on</strong>s (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance facial expressi<strong>on</strong>s);- by asking the participant to self-assess his / her feelings after the stimulati<strong>on</strong>.The a-priori annotati<strong>on</strong> can be d<strong>on</strong>e arbitrarily <str<strong>on</strong>g>for</str<strong>on</strong>g> instance according to the judgment of theexperimenter. However this method is not recommended since it does not take into account thevariability of judgments that can be observed in a populati<strong>on</strong>. Another possibility is to ask a largepopulati<strong>on</strong> of pers<strong>on</strong>s to extensively evaluate the stimuli. Each stimulus can then be associated tothe most frequent label or to the average of judgments if emoti<strong>on</strong>s were evaluated in a c<strong>on</strong>tinuousspace such as the valence-arousal space. For instance, each of the IAPS images [54] is providedwith the mean <strong>and</strong> the st<strong>and</strong>ard deviati<strong>on</strong> of the valence-arousal judgments of 800 pers<strong>on</strong>s.However this method is still not optimal since the participants can have various emoti<strong>on</strong>alreacti<strong>on</strong>s under the same stimulus, due <str<strong>on</strong>g>for</str<strong>on</strong>g> instance to differences in past experience. As anexample, an image of some<strong>on</strong>e skiing can elicit pleasure but it can also elicit negative emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g>a participant that had a bad experience <strong>on</strong> skis. To alleviate this problem, it is possible to ask toeach participant to select <strong>and</strong> evaluate images be<str<strong>on</strong>g>for</str<strong>on</strong>g>e the experiment. By using this method theelicited emoti<strong>on</strong>s are known in advance. The problem is then that the stimuli will be known by theparticipants <strong>and</strong> thus elicit different emoti<strong>on</strong>al resp<strong>on</strong>se during the experiment (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance lessintense resp<strong>on</strong>ses).Analyzing the emoti<strong>on</strong>al behavior <strong>and</strong> expressi<strong>on</strong>s of the participant is a possible method todetermine the elicited emoti<strong>on</strong>. However this method requires that at least <strong>on</strong>e expert (apsychologist <str<strong>on</strong>g>for</str<strong>on</strong>g> instance) is present during the experiment. In order to avoid bias in the estimatedemoti<strong>on</strong>s, it is even better to group the judgments of many experts, raising the questi<strong>on</strong> of thecombinati<strong>on</strong> of different judgments. The participant should be free to fully express his / heremoti<strong>on</strong>s, which is not always the case. For instance, when physiological signals are recordedmovements are generally limited to avoid noise in the signals. Finally, if a model is trained fromthis type of ground-truth then it will be able to detect the expressi<strong>on</strong> of the emoti<strong>on</strong> <strong>and</strong> not theother factors involved in emoti<strong>on</strong>s such as the subjective feelings (see Secti<strong>on</strong> 2.1.3.b).72


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentPresenting the stimuli to the participants <strong>and</strong> then asking them to self-assess their feeling is alsoan alternative to the two precedent annotati<strong>on</strong> methods. This can be d<strong>on</strong>e by asking theparticipants to fill in questi<strong>on</strong>naires, to give emoti<strong>on</strong>al labels <strong>and</strong> to evaluate their emoti<strong>on</strong>s in thevalence-arousal space. While this method alleviates the problems related to the subjectivity ofemoti<strong>on</strong>s it relies <strong>on</strong> the assumpti<strong>on</strong> that the participants are reliable experts in evaluating theirown feelings. This latter statement can be discussed because of the following issues. Firstly, theparticipant will determine the elicited emoti<strong>on</strong>s mostly based <strong>on</strong> his / her subjective feeling whichis <strong>on</strong>ly <strong>on</strong>e of the factors involved in the emoti<strong>on</strong> (see Secti<strong>on</strong> 2.1.3.b). Sec<strong>on</strong>dly,misunderst<strong>and</strong>ing of the material used <str<strong>on</strong>g>for</str<strong>on</strong>g> self-assessment could lead to wr<strong>on</strong>g annotati<strong>on</strong>s. Asstated in Secti<strong>on</strong> 2.1.4, emoti<strong>on</strong>al words can have different meanings across pers<strong>on</strong>s <strong>and</strong> cultures.Moreover, representing an emoti<strong>on</strong> in the valence-arousal space is not straight<str<strong>on</strong>g>for</str<strong>on</strong>g>ward <strong>and</strong> it ism<strong>and</strong>atory to provide explanati<strong>on</strong>s be<str<strong>on</strong>g>for</str<strong>on</strong>g>e some<strong>on</strong>e can use this space. Thirdly, a participant canhide his / her true felt emoti<strong>on</strong>al state because of social rules. For instance, a man can hesitate toreport high arousal while watching a picture of a nude man.The best way to annotate data is certainly to combine the different annotati<strong>on</strong> methods describedabove. An example can be to combine annotati<strong>on</strong>s from experts that evaluate an emoti<strong>on</strong>al statebased <strong>on</strong> the analysis of several emoti<strong>on</strong>al cues (facial expressi<strong>on</strong>s, speech, behavior,physiological signals etc.) together with self-reported measures of emoti<strong>on</strong>s. This would enableto determine a ground-truth based <strong>on</strong> several comp<strong>on</strong>ents involved in emoti<strong>on</strong>al processes.However it is not clear how the fusi<strong>on</strong> of the different annotati<strong>on</strong> should be per<str<strong>on</strong>g>for</str<strong>on</strong>g>med.In this study both self-generated <strong>and</strong> stimuli-based emoti<strong>on</strong>s were elicited in different protocols.For the protocol where emoti<strong>on</strong>s were self-generated, the trials were directly annotated with thecorresp<strong>on</strong>ding emoti<strong>on</strong>al state as explained above. For the protocols where emoti<strong>on</strong>s wereelicited using stimuli two of the annotati<strong>on</strong> methods described above were employed: a-prioriannotati<strong>on</strong> <strong>and</strong> self-assessment. The effectiveness of those methods was thus estimated based <strong>on</strong>the accuracy of the emoti<strong>on</strong> assessment.The above paragraphs detail how to collect a ground-truth without any assumpti<strong>on</strong> <strong>on</strong> the type ofannotati<strong>on</strong>s that are collected. Those annotati<strong>on</strong>s could be c<strong>on</strong>tinuous (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance by using thevalence-arousal space) or discrete (using emoti<strong>on</strong>al labels such as fear <strong>and</strong> anger). Essentiallybecause of its generality but also <str<strong>on</strong>g>for</str<strong>on</strong>g> other reas<strong>on</strong>s detailed in Secti<strong>on</strong> 2.1.4 the valence-arousalspace was chosen as a representati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>al states. Accurately determining a point in thisspace based <strong>on</strong>ly <strong>on</strong> physiological features is a difficult task. For this reas<strong>on</strong> we preferred to takea first step by defining valence-arousal classes of interest. Thus each value y i of the y vector cantake values in a set of emoti<strong>on</strong>al labels Y={ 1 ,…, c }, where C is the number of classes.73


Chapter 4Depending <strong>on</strong> the protocol, the valence-arousal space was thus divided in different regi<strong>on</strong>s <strong>and</strong>each regi<strong>on</strong> associated to a target label. For instance, three regi<strong>on</strong>s of interest (C=3) can bedefined by segmenting the valence-arousal space in calm, excited-positive <strong>and</strong> excited negativeareas. Another possibility is to segment the space in two areas (C=2) such as calm vs. excited orpositive vs. negative areas. The segmentati<strong>on</strong> can be d<strong>on</strong>e a-priori, the participants thus annotatethe emoti<strong>on</strong>s accordingly to the classes defined; or a-posteriori from c<strong>on</strong>tinuous annotati<strong>on</strong>sgathered during the protocol. All those different possibilities give rise to different y ground-truthvectors <strong>and</strong> thus corresp<strong>on</strong>ds to different <str<strong>on</strong>g>for</str<strong>on</strong>g>mulati<strong>on</strong>s of the emoti<strong>on</strong> assessment problem thatare called classificati<strong>on</strong> schemes. Several classificati<strong>on</strong> schemes were studied <str<strong>on</strong>g>for</str<strong>on</strong>g> the datagathered from each protocol <strong>and</strong> they will be detailed in the appropriate chapters.4.1.2 Validati<strong>on</strong> strategiesFrom the ground-truth acquired according to the methods presented in Secti<strong>on</strong> 4.1.1, the emoti<strong>on</strong>assessment task is defined as supervised classificati<strong>on</strong>. It is supervised because a ground-truth isavailable to learn a model (the y i values) <strong>and</strong> it is classificati<strong>on</strong> because the goal is to retrieveemoti<strong>on</strong>al classes of interest y ˆi. In this case, the methods usable <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment originatefrom the pattern recogniti<strong>on</strong> <strong>and</strong> machine learning fields [119, 120].When training classifiers overfitting can occur when the obtained model perfectly fits the datafrom which it is learned but per<str<strong>on</strong>g>for</str<strong>on</strong>g>ms poorly <strong>on</strong> new unseen data [119]. In order to c<strong>on</strong>trol <str<strong>on</strong>g>for</str<strong>on</strong>g> thegeneralizati<strong>on</strong> capability of the model, it is thus important to test the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of a learnedmodel (the classifier) <strong>on</strong> a different dataset than the <strong>on</strong>e used <str<strong>on</strong>g>for</str<strong>on</strong>g> learning. Validati<strong>on</strong> strategiesc<strong>on</strong>sist in segmenting the data in two sets: a training set from which the model is learned <strong>and</strong> atest set <strong>on</strong> which the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of the model is tested (Figure 4.1).f 1.............f NFy 1.............y 2yDivisi<strong>on</strong>in 2 setsTraining setParametersLearningClassificati<strong>on</strong>Computati<strong>on</strong>al ModelTestingTest setŷAFigure 4.1. Validati<strong>on</strong> scheme <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong>, where ŷ is the vector of the classes estimated by model <str<strong>on</strong>g>for</str<strong>on</strong>g> thetest set, A is the accuracy.In this study, the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of a model was tested by using the following measure of accuracy:74


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentANNc (4.1)where N t is the number of samples in the test set <strong>and</strong> N c is the number of test samples correctlyclassified (test samples where yˆi y i). A c<strong>on</strong>fusi<strong>on</strong> matrix (Table 4.1) will also be used todetermine how the samples are classified in the different classes. A c<strong>on</strong>fusi<strong>on</strong> matrix gives thepercentage of samples bel<strong>on</strong>ging to class i <strong>and</strong> classified as class j .Groundtruth labelsytEstimated labels ŷ 1 . j . c 1 P 1,1 P 1,c. i. c P c,1 P c,cTable 4.1. A c<strong>on</strong>fusi<strong>on</strong> matrix, P i,j is the percentage of samples bel<strong>on</strong>ging to class i <strong>and</strong> classified as class j .The accuracy A can be retrieved from the c<strong>on</strong>fusi<strong>on</strong> matrix by summing its diag<strong>on</strong>al elements P i,iweighted by the prior probability p( i ) of occurrence of the class i :P i,jcA p( ) P(4.2)i1For the model to correctly represent the data, it is important that the training set c<strong>on</strong>tains enoughsamples (or instances). On the other h<strong>and</strong> it also important that the test set c<strong>on</strong>tains enoughsamples to avoid a noisy estimate of the model per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance. This can be problematic because itoften occurs that the amount of collected data is limited in practice. This is particularly true in ourcase since the number of emoti<strong>on</strong>al stimulati<strong>on</strong>s is limited by the durati<strong>on</strong> of the protocols whichshould not be too l<strong>on</strong>g to avoid participant fatigue as well as elicitati<strong>on</strong> of undesired emoti<strong>on</strong>s.Cross-validati<strong>on</strong> methods help to solve this problem by splitting the data in different training /test sets so that each sample will be used at least <strong>on</strong>ce <str<strong>on</strong>g>for</str<strong>on</strong>g> training <strong>and</strong> <strong>on</strong>ce <str<strong>on</strong>g>for</str<strong>on</strong>g> testing.The two well known cross-validati<strong>on</strong> methods are the k-fold <strong>and</strong> the leave-<strong>on</strong>e-out [136]. In thek-fold cross-validati<strong>on</strong>, the data is split in k folds c<strong>on</strong>taining the same amount of samples.Generally the folds are determined so that the prior probability p( i ) of observing each class i isthe same <str<strong>on</strong>g>for</str<strong>on</strong>g> each fold. Each fold is then used in turn as the test set <strong>and</strong> a model is learned fromthe remaining k-1 folds. By using this method k accuracies are obtained from the k test sets sothat it is possible to compute the average accuracy <strong>and</strong> its st<strong>and</strong>ard deviati<strong>on</strong>. The leave-<strong>on</strong>e-outcross-validati<strong>on</strong> is similar to the k-fold cross-validati<strong>on</strong> except that the size of the test set isalways 1. Thus N models are tested in turn <strong>on</strong> each sample <strong>and</strong> learned from the N-1 remainingsamples of the database. The advantage of this cross-validati<strong>on</strong> method is that it provides theiii ,75


Chapter 4maximum possible size <str<strong>on</strong>g>for</str<strong>on</strong>g> the training set which generally helps to find a better model,especially in the case where few samples are available in the database. On the other h<strong>and</strong> <strong>on</strong>ly theaverage accuracy can be computed reliably since the test set c<strong>on</strong>tains <strong>on</strong>ly <strong>on</strong>e sample (theaccuracy is thus either 0 or 1).When designing a general computati<strong>on</strong>al model <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment (i.e. a model that is notpers<strong>on</strong> dependent but can be used to assess emoti<strong>on</strong>s of any<strong>on</strong>e) based <strong>on</strong> physiological featuresit is important to take into account the high variability that can be observed in physiologicalreacti<strong>on</strong>s. To c<strong>on</strong>trol the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of such a model it should be tested <strong>on</strong> physiological data ofpers<strong>on</strong>s whose features were not used <str<strong>on</strong>g>for</str<strong>on</strong>g> the learning of the model. For this reas<strong>on</strong> theparticipant cross-validati<strong>on</strong> method was proposed. The database was segmented in folds whereeach fold c<strong>on</strong>tains the samples computed from the physiological signals of a single participant.Then the classificati<strong>on</strong> per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance was computed similarly to the k-fold cross-validati<strong>on</strong>, byusing each fold in its turn as the test set. This method allows testing the classificati<strong>on</strong>per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance as in a “real-case” where the emoti<strong>on</strong>s of a user would be assessed by using a modeldefined from the physiological activity of other pers<strong>on</strong>s.4.1.3 ClassifiersSecti<strong>on</strong> 4.1.2 detailed how to determine the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of a classifier. This secti<strong>on</strong> will describethe different classifiers used in this study, all being part of the pattern recogniti<strong>on</strong> <strong>and</strong> themachine learning fields [119, 120]. For most classificati<strong>on</strong> algorithms, it is important that thefeatures be normalized (i.e. bel<strong>on</strong>gs to the same range of value). This normalizati<strong>on</strong> was appliedat each cross-validati<strong>on</strong> step by whithening each feature using mean <strong>and</strong> st<strong>and</strong>ard deviati<strong>on</strong>computed from the training set.a. Naïve BayesSeveral classifiers rely <strong>on</strong> the Bayes’ rule to find the most probable class in which a samplerepresented by its feature vector f should be classified. This is d<strong>on</strong>e by attributing the class i thatmaximize the posterior probability p( i | f) to the estimated label ŷ . According to the Bayes’rule:p( | f)iCi1p( f | ) p( )ip( f | ) p( )iii(4.3)One of the advantages of this type of classifier is that it is able to output the posterior probabilityp( i | f) that a sample bel<strong>on</strong>g to a given class i . Notice that it is enough to find the maximumvalue of the numerator to maximize p( i | f) since the denominator has the same value <str<strong>on</strong>g>for</str<strong>on</strong>g> any76


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentclass i . The main differences between Bayesian classifiers is the way by which the c<strong>on</strong>diti<strong>on</strong>alprobabilities p(f | i ) are estimated.For the Naïve-Bayes classifier the assumpti<strong>on</strong> of c<strong>on</strong>diti<strong>on</strong>al independence of the features given i is made:pF( f | ) p( f | )(4.4)i j ij1where F is the number of features in the feature vector f. In this study, the c<strong>on</strong>diti<strong>on</strong>al probabilityp(f j | i ) of a feature j was estimated by quantizing the features in 10 bins of equal sizes <strong>and</strong>computing the associated c<strong>on</strong>diti<strong>on</strong>al probability mass functi<strong>on</strong> from the training set. The priorprobability p( i ) could also be computed from the training set; it would however be biased by thestimuli presented <str<strong>on</strong>g>for</str<strong>on</strong>g> each class in the protocol. For instance, if the aim of the classifier is todistinguish between calm <strong>and</strong> excited emoti<strong>on</strong>al states <strong>and</strong> more excited stimulus were presentedto the participants then the prior probability would be higher <str<strong>on</strong>g>for</str<strong>on</strong>g> the excited class. While this iscoherent in this particular protocol it does not have any meaning in real applicati<strong>on</strong>s since there isnothing that guarantees the higher occurrence of excited states in this case. For this reas<strong>on</strong> theprior probability p( i ) was set to 1/C under the assumpti<strong>on</strong> of equiprobability of the classes.b. Discriminant analysisTwo discriminant analysis methods, namely the linear discriminant analysis (LDA) <strong>and</strong> theQuadratic discriminant analysis (QDA) are used in this study. Both are based <strong>on</strong> the Bayes rule tofind the class with the highest posterior probability p( i | f) [119]. For this purpose the followingg i discriminant functi<strong>on</strong>s are defined:g ( f) ln( p( f | ) p( ))(4.5)i i iFinding the class i with the highest g i value is then similar to finding the class that maximizesthe numerator of equati<strong>on</strong> 4.3 Under the assumpti<strong>on</strong> that the c<strong>on</strong>diti<strong>on</strong>al distributi<strong>on</strong>s p(f | i ) areGaussians with different means µ i <strong>and</strong> covariance matrices i this rule automatically defines a(hyper-)quadratic decisi<strong>on</strong> boundary (hence the name QDA <str<strong>on</strong>g>for</str<strong>on</strong>g> the associated classifier):g1 F 1f f f (4.6)2 2 21Ti( ) ( i) i( i) ln 2 lniln p( i)where T <strong>and</strong> |.| respectively st<strong>and</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> the transpose <strong>and</strong> determinant operators. Vectors µ i <strong>and</strong>matrices i are computed from the training set. In the case where i j,ij the boundarybecomes linear, yielding an LDA classifier. With the LDA it is sufficient to compute a single77


Chapter 4covariance matrix from the complete training set without distincti<strong>on</strong> between classes. Similarlyto the Naïve Bayes classifier, the prior probability p( i ) was defined as 1/C.In the case where the size of the feature space F is large <strong>and</strong> the number of samples available <str<strong>on</strong>g>for</str<strong>on</strong>g>learning is small, the discriminant analysis can fall in the singularity problem where thematrix is not invertible. In this case we used the diag<strong>on</strong>alized versi<strong>on</strong> where covariance matricesare assumed to be diag<strong>on</strong>al, c<strong>on</strong>taining the variances of the features. Notice that in this case thediscriminant analysis is a Naïve Bayes classifier with a c<strong>on</strong>diti<strong>on</strong>al independent Gaussianassumpti<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the p(f j | i ) distributi<strong>on</strong>s. The Matlab statistics toolbox (v. 5.0.1) implementati<strong>on</strong>of those algorithms was used in this study.c. Support Vector Machines (SVM’s)A SVM [120, 137] is a two class classifier (C=2) using a linear model of the <str<strong>on</strong>g>for</str<strong>on</strong>g>m:h() f w() f T b(4.7)where a feature vector f is estimated as being from class 1 if h(f)0. The functi<strong>on</strong> projects a feature vector in another feature space, generally of higher dimensi<strong>on</strong>ality,thus allowing <str<strong>on</strong>g>for</str<strong>on</strong>g> n<strong>on</strong> linear separati<strong>on</strong> of the data in the original feature space. In order to findthe model weights w <strong>and</strong> b, an SVM tries to maximize the distance between the decisi<strong>on</strong> surfacecreated by the h functi<strong>on</strong> <strong>and</strong> a margin to this surface as well as to minimize the error <strong>on</strong> thetraining set. The trade-off between margin maximizati<strong>on</strong> <strong>and</strong> the training error minimizati<strong>on</strong> isc<strong>on</strong>trolled by a parameter C SVM that was empirically set to 1 in this study. The advantage ofSVM's is that they minimize an upper bound <strong>on</strong> the expected risk rather than <strong>on</strong>ly the error <strong>on</strong> thetraining data, thus enabling good generalizati<strong>on</strong> even <str<strong>on</strong>g>for</str<strong>on</strong>g> undersampled datasets, as well asinteresting per<str<strong>on</strong>g>for</str<strong>on</strong>g>mances in high dimensi<strong>on</strong>al feature spaces [138]. Moreover, they provide sparsesoluti<strong>on</strong>s where not all of the data points are used <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong>.The SVM optimizati<strong>on</strong> problem can be expressed in a dual <str<strong>on</strong>g>for</str<strong>on</strong>g>m [120, 137], where a kernelfuncti<strong>on</strong> k(, f f) ()( f f ) T is introduced between two samples f <strong>and</strong> f´. In this new <str<strong>on</strong>g>for</str<strong>on</strong>g>mulati<strong>on</strong>,the decisi<strong>on</strong> boundary becomes a functi<strong>on</strong> of <strong>on</strong>ly some of the data points called the supportvectors. In this study, both linear <strong>and</strong> radial basis functi<strong>on</strong> (RBF) kernels were used:linearTk ( f, f) ff . (4.8)1i 2f fRBFk ( ff , )e (4.9)78


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentwhere ||.|| is the norm operator. In the case of RBF kernels, the size of the kernel was chosen byapplying a 5-fold cross-validati<strong>on</strong> procedure <strong>on</strong> the training set <strong>and</strong> finding the yielding the bestaccuracy. The tested values bel<strong>on</strong>ged to the 5.10 -3 to 5.10 -1 range with a step of 5.10 -3 .There are two drawbacks to the use of SVM’s as classifiers: they are intrinsically <strong>on</strong>ly two-classclassifiers <strong>and</strong> their output is uncalibrated so that it is not directly usable as a c<strong>on</strong>fidence value inthe case <strong>on</strong>e wants to combine outputs of different classifiers or modalities. In this study the firstpoint was addressed by using the <strong>on</strong>e-versus-<strong>on</strong>e approach where C(C-1)/2 classifiers are trained<strong>on</strong> each possible pair of classes. The class associated to a test sample is the <strong>on</strong>e that receives thehighest number of votes from the C(C-1)/2 classifiers.p(1 | h)a) h functi<strong>on</strong> valueb)h functi<strong>on</strong> valueFigure 4.2. Obtaining posterior probabilities p( i | h) from SVM outputs. a) Histograms representing thedistributi<strong>on</strong>s of the SVM output <str<strong>on</strong>g>for</str<strong>on</strong>g> two classes. b) Posterior probabilities estimates from the Bayes rulesapplied <strong>on</strong> the histogram of a) <strong>and</strong> from the sigmoid fit proposed by Platt [139].For the sec<strong>on</strong>d point, Platt (2000) proposed to model the probability p( 1 | h) of being in the firstof the two classes knowing the output value h of the SVM. As can be seen in Figure 4.2 thiscould be d<strong>on</strong>e by applying the Bayes rule. The discrete posterior probability plotted in Figure 4.2approximately follows a sigmoid curve; this is why Platt proposed to model those probabilities byusing:1p( 1| h) 1 exp( h )(4.10)where the <strong>and</strong> values are found by the algorithm proposed in [139] <strong>and</strong> improved in [140].Figure 4.2 shows the result of the sigmoid curve fitting. C<strong>on</strong>cretely, the h values were obtainedfrom a 5-fold cross-validati<strong>on</strong> <strong>on</strong> the training set <strong>and</strong> the parameters <strong>and</strong> were determinedfrom those h values. The posterior probabilities of the test samples were then computed usingequati<strong>on</strong> 4.10. Finally, to compute the posterior probabilities p( i | h) when there are more thantwo classes to separate, the soluti<strong>on</strong> proposed in [141] was employed. The libSVM [142] Matlabtoolbox was used as an implementati<strong>on</strong> of the SVM <strong>and</strong> probabilistic SVM algorithms.79


Chapter 4d. Relevance Vector Machines (RVM’s)RVM’s [143] are algorithms that have the same functi<strong>on</strong>al <str<strong>on</strong>g>for</str<strong>on</strong>g>m as SVM's but embedded in aBayesian learning framework. They have been shown to provide results similar to SVM's withgenerally sparser soluti<strong>on</strong>s. They have the advantage that they directly give an estimati<strong>on</strong> of theposterior probability of having class i .RVM’s try to maximize the likelihood functi<strong>on</strong> of the training set using a linear model includingkernels. The main difference with classical probabilistic discriminative models is that a differentprior is applied <strong>on</strong> each weight thus leading to sparse soluti<strong>on</strong>s that should generalize well. Forall the following studies, the multiclass RVM versi<strong>on</strong> presented in [144] was used.4.2 Feature selecti<strong>on</strong>The features extracted from the physiological signals, especially EEG signals, were sometimes ofhigh dimensi<strong>on</strong>ality. For instance the f EEG_STFT feature vector presented in secti<strong>on</strong> 3.4.2.a wouldc<strong>on</strong>tain 16704 features <str<strong>on</strong>g>for</str<strong>on</strong>g> an EEG recorded over a period of 7.5 sec<strong>on</strong>ds with 64 electrodes.Although increasing the number of features in a database should theoretically decreaseclassificati<strong>on</strong> error, this is not the case in practice because models of classifiers, or parameters ofdistributi<strong>on</strong>s, are estimated from a training set of limited size. Having fewer samples thanfeatures <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong>, as could be the case when the feature space is of high dimensi<strong>on</strong>ality,is known as the undersampled or singularity problem [138, 145]. To alleviate this problem, <strong>on</strong>ecan resort to feature space reducti<strong>on</strong> techniques. Reducing the size of a feature space can be d<strong>on</strong>eeither by finding a functi<strong>on</strong> that projects the data points in a space of lower dimensi<strong>on</strong>ality(generally preserving or improving the discriminability between classes) [119, 146] or bydiscarding the features that are not of interest <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> [147-149]. In this studyalgorithms from the sec<strong>on</strong>d approach, called feature selecti<strong>on</strong> algorithms, were implemented.Two different approaches are generally c<strong>on</strong>sidered to deal with feature selecti<strong>on</strong> [147]: the filterapproach where the quality of each feature is estimated independently of the classificati<strong>on</strong>scheme, generally using statistical measures; <strong>and</strong> the wrapper approach that relies <strong>on</strong> theclassificati<strong>on</strong> algorithm to evaluate feature subsets as well as <strong>on</strong> heuristic methods to find the bestsubset. Although the sec<strong>on</strong>d approach can yield better results, it suffers from an importantcomputati<strong>on</strong>al cost, especially <str<strong>on</strong>g>for</str<strong>on</strong>g> very high dimensi<strong>on</strong>al space. There are also classificati<strong>on</strong>algorithms that per<str<strong>on</strong>g>for</str<strong>on</strong>g>m embedded feature selecti<strong>on</strong> such as the Adaboost [150]. The SVM alsoc<strong>on</strong>tains a regularizati<strong>on</strong> term in its error functi<strong>on</strong> that keeps the weights w low <strong>and</strong> thus providesa <str<strong>on</strong>g>for</str<strong>on</strong>g>m of feature selecti<strong>on</strong>.In this study, feature selecti<strong>on</strong> algorithms were always applied <strong>on</strong> the training set to determine theselected subset of features. The classifiers were then trained <strong>on</strong> this selected subset <strong>and</strong> the80


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentresulting model was applied <strong>on</strong> the test features subset. This secti<strong>on</strong> describes the filter <strong>and</strong>wrapper algorithms used <str<strong>on</strong>g>for</str<strong>on</strong>g> feature selecti<strong>on</strong>.4.2.1 Filter methodsa. ANOVAThe ANOVA (ANalysis Of VAriance) is a statistical test that estimates if a difference in meanobserved between several groups <str<strong>on</strong>g>for</str<strong>on</strong>g> a given variable is significant. Two values can be computedfrom an ANOVA test: the F value that gives the magnitude of the mean difference <strong>and</strong> the pvalue that gives the probability of making an error by assuming a difference in mean. If the pvalue is low (near to 0) than this reflects the str<strong>on</strong>g belief that at least <strong>on</strong>e group mean is differentfrom the others.p(fi | i)f if iFigure 4.3. Different possible distributi<strong>on</strong>s of a feature value <str<strong>on</strong>g>for</str<strong>on</strong>g> a 3 classes scenario (green, red <strong>and</strong> blackclasses). (left) The feature is relevant since it is usefull to distinguish the green class from the others (low pvalue). (right) A n<strong>on</strong> relevant feature (high p value).Under the assumpti<strong>on</strong> that the c<strong>on</strong>diti<strong>on</strong>al distributi<strong>on</strong>s p(f i | i ) are Gaussians with similarvariances than a feature is relevant <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> if the means of those distributi<strong>on</strong>s aredifferent <str<strong>on</strong>g>for</str<strong>on</strong>g> at least <strong>on</strong>e of the classes (Figure 4.3). Thus an ANOVA test was applied <strong>on</strong> eachfeature with the groups defined by the classes <strong>and</strong> the resulting p values were used to determine ifa feature was relevant. A feature was c<strong>on</strong>sidered to be relevant if <strong>and</strong> <strong>on</strong>ly if p < ANOVA , with ANOVA set to 0.1 which corresp<strong>on</strong>ds to an error probability of 10%. All n<strong>on</strong>-relevant featureswere discarded. It was thus not possible to a-priori choose an exact number of features to beselected.b. Fisher criteri<strong>on</strong>The Fisher projecti<strong>on</strong> [119, 146] is a well known algorithm that projects the samples of a givenfeature space into a subspace of lower dimensi<strong>on</strong>ality having the highest linear discriminati<strong>on</strong>between the classes according to the Fisher criteri<strong>on</strong>. However, the choice of the dimensi<strong>on</strong> F s ofthe subspace is limited by the number of classes (F s < C). In order to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m feature selecti<strong>on</strong>instead of projecti<strong>on</strong> <strong>and</strong> to be free to choose the size of the resulting feature space, the Fishercriteri<strong>on</strong> used in the Fisher projecti<strong>on</strong> algorithm was applied independently to each feature. For81


Chapter 4this purpose, vectors f ic<strong>on</strong>taining all the samples <str<strong>on</strong>g>for</str<strong>on</strong>g> each feature f i were c<strong>on</strong>structed from thefeature set F (a vector f iis a column of F). It is then possible to apply the following Fishercriteri<strong>on</strong> <strong>on</strong> such a vector:J()fCj1CN ( m m)jj1fDjj2( fm)j2(4.11)where m <strong>and</strong> m j are respectively the mean of the vector f <strong>and</strong> the mean of this vector <str<strong>on</strong>g>for</str<strong>on</strong>g> samplesbel<strong>on</strong>ging to class j, N j is the number of samples bel<strong>on</strong>ging to class j <strong>and</strong> D j represents theensemble of values from f that bel<strong>on</strong>gs to class j.From the numerator of Equati<strong>on</strong> 4.11 it can be seen that the larger the distance between thefeature means of each classes, the higher the value of J. On the other h<strong>and</strong>, the denominatorrepresents the average variability of the feature across the classes <strong>and</strong> the smaller the better. Thusthis criteri<strong>on</strong> will have high values <str<strong>on</strong>g>for</str<strong>on</strong>g> features that have different mean <str<strong>on</strong>g>for</str<strong>on</strong>g> each class <strong>and</strong> a lowaverage intra-class variance which is very similar to what the ANOVA test does.To filter out features, the Fisher criteri<strong>on</strong> J was computed <str<strong>on</strong>g>for</str<strong>on</strong>g> each feature <strong>and</strong> the features werethen ranked by decreasing order of J. Finally, the F s features with the highest J value were kept<strong>and</strong> the others removed. This algorithm thus allows selecting the size F s of the resulting featurespace.c. Fast Correlati<strong>on</strong> <str<strong>on</strong>g>Based</str<strong>on</strong>g> Filter (FCBF)The FCBF algorithm [149] was proposed as a filter method that selects features according to theirrelevance to the class c<strong>on</strong>cept <strong>and</strong> removes those that are redundant. By removing redundantfeatures the algorithm reduces the size of the selected subset of features without changing itsdisciminant capacity. The absolute value of the linear correlati<strong>on</strong> measure was used to evaluateboth the relevance <strong>and</strong> redundancy of a feature:Corr( , )m( x x)( z z)i ii1xz (4.12)mm2 2 ( xix) ( ziz)i1 j1where x, z are two vectors of size m respectively c<strong>on</strong>taining x i <strong>and</strong> z i values with means x <strong>and</strong> z .82


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentFrom the feature set F, vectors f iwere c<strong>on</strong>structed. These vectors c<strong>on</strong>tains all the samples ofeach feature f i . The algorithm then selected a feature subset in two steps:- removal of each irrelevant feature f i where Corr( f , y) ;- removal of redundant features; a feature f i being c<strong>on</strong>sidered as redundant with respect toanother feature f j if Corr( f , f ) Corr( f , y)<strong>and</strong> Corr( f , y) Corr( f , y).i j jThe first step removes features that have low correlati<strong>on</strong>s with the classes. In the sec<strong>on</strong>d step thepairs of features that are highly correlated are identified <strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> each feature pair the feature thathas the lowest correlati<strong>on</strong> with the class is removed.iiFCBFjAs <str<strong>on</strong>g>for</str<strong>on</strong>g> the ANOVA feature selecti<strong>on</strong>, the number of selected features is determined by the FCBFparameter but the exact number of features can not be chosen explicitly. The possible values <str<strong>on</strong>g>for</str<strong>on</strong>g>the FCBF parameter range from 0 to 1. With FCBF = 0 all the features are selected in the first stepbut redundant features are removed in the sec<strong>on</strong>d. When FCBF is increased the number ofselected features decreases with the minimum number of selected features being reached when FCBF = 1 (<strong>on</strong>ly the features that are fully correlated with the classes are kept).4.2.2 The SFFS wrapper methodIn most of the filter algorithms, the relevance of a feature is estimated independently from theother features. However it can happen that a feature is not relevant by its own but is highlyrelevant when combined with another feature; this is called feature interacti<strong>on</strong> [151, 152].Wrapper algorithms have the advantage that a feature is added or removed from a feature setbased <strong>on</strong> the analysis of the newly generated feature set. As a c<strong>on</strong>sequence, those algorithms areable to remove redundant features but also to take into account the interacti<strong>on</strong> between features.Moreover, the wrapper algorithms generally use the predictive accuracy of the classifier used <str<strong>on</strong>g>for</str<strong>on</strong>g>final classificati<strong>on</strong> to evaluate the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of a feature subset. Thus the features are selectedin accordance with the intrinsic properties of the classifier. However the main drawback of thosealgorithms is that they are computati<strong>on</strong>ally intensive since they require the evaluati<strong>on</strong> of anumber of feature subsets that is generally higher than the number of features.The Sequential Floating Forward Selecti<strong>on</strong> (SFFS) [148] was implemented as a wrapperalgorithm if the number of features was sufficiently low. This algorithm starts with an emptyfeature set. At each iterati<strong>on</strong> of the algorithm a <str<strong>on</strong>g>for</str<strong>on</strong>g>ward step is taken followed by a number ofbackward steps. A <str<strong>on</strong>g>for</str<strong>on</strong>g>ward step c<strong>on</strong>sists in adding to the current feature set the feature thatmaximizes the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of the newly created feature set. A backward step c<strong>on</strong>sists inremoving a feature from the current feature set <strong>on</strong>ly if doing so improves the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance. Thislast step is taken as l<strong>on</strong>g as removing a feature improves the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance. This method searches83


Chapter 4the feature space more deeply than by using <strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g>ward or backward algorithms [148]. TheSFFS algorithm stops when the number of features in the subset is equal to the number ofrequested features F SFFS <strong>and</strong> no backward step is taken. The best subset of selected features isthen returned which can be of size 1 to F SFFS . In this study the classificati<strong>on</strong> accuracy, computed<strong>on</strong> the samples of the training set, was used as the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance measure. The classifier used toevaluate a feature subset was the same as the <strong>on</strong>e used to generate the final classificati<strong>on</strong> model.For instance, if the SVM classifier was used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment then it was also used <str<strong>on</strong>g>for</str<strong>on</strong>g>feature selecti<strong>on</strong>.4.3 Fusi<strong>on</strong>As stated in Secti<strong>on</strong>s 1.1.2 <strong>and</strong> 2.1.3.b emoti<strong>on</strong> elicitati<strong>on</strong> is a multimodal process that involvesseveral comp<strong>on</strong>ents of the organism. As a c<strong>on</strong>sequence, emoti<strong>on</strong>s are expressed through severalchannels, giving rise to many emoti<strong>on</strong>al cues that can be recorded by different sensors. Since thein<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> recorded by those sensors can represent the activity of the different comp<strong>on</strong>entsinvolved in emoti<strong>on</strong>al processes, combining the in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> obtained from those sensors canimprove the reliability of emoti<strong>on</strong> assessment. The combinati<strong>on</strong> of multi-sensor in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> isknown as fusi<strong>on</strong>. In this study, several sensors described in Secti<strong>on</strong> 3.1 were used to m<strong>on</strong>itor theactivity of both the peripheral <strong>and</strong> the central nervous system. Moreover, different types offeatures were also computed <str<strong>on</strong>g>for</str<strong>on</strong>g> each sensor. This secti<strong>on</strong> describes how the different in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>obtained from the sensors were fused <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of emoti<strong>on</strong> classificati<strong>on</strong>.When the goal is to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m classificati<strong>on</strong>, in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> fusi<strong>on</strong> can be d<strong>on</strong>e at different levelsincluding the sensor data, the feature <strong>and</strong> the classifier levels [153, 154]. In this study theper<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of fusi<strong>on</strong> was evaluated at the feature <strong>and</strong> classifier level. In those cases, theproblem could be <str<strong>on</strong>g>for</str<strong>on</strong>g>mulated as having a dataset c<strong>on</strong>taining several feature sets F j (instead of <strong>on</strong>eas presented in Secti<strong>on</strong> 4.1) <strong>and</strong> the associated label vector y. Each of those feature sets canc<strong>on</strong>tain:- the features extracted from the signals produced by a given sensor (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance all thefeatures extracted from the GSR signals or all the features extracted from the EEGsignals);- the features extracted from the signals corresp<strong>on</strong>ding to <strong>on</strong>e of the two parts of thenervous system (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance the features extracted from the PNS or those extracted fromthe CNS);- the different features extracted from the signals of a unique sensor (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance a featureset c<strong>on</strong>taining the MI features extracted from EEG signals <strong>and</strong> another <strong>on</strong>e c<strong>on</strong>taining theSTFT features extracted from the same signals).84


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentUnder this <str<strong>on</strong>g>for</str<strong>on</strong>g>mulati<strong>on</strong>, fusi<strong>on</strong> at the feature levels c<strong>on</strong>sists in combining the features sets F jbe<str<strong>on</strong>g>for</str<strong>on</strong>g>e classificati<strong>on</strong>, while fusi<strong>on</strong> at the classifier level c<strong>on</strong>sists in classifying each feature setindependently <strong>and</strong> combine the classificati<strong>on</strong> results in a sec<strong>on</strong>d step.4.3.1 Feature levelFusi<strong>on</strong> at the feature level can be d<strong>on</strong>e by c<strong>on</strong>catenating the feature sets or by linearly or n<strong>on</strong>linearlycombining the features [153]. C<strong>on</strong>catenati<strong>on</strong> of features from the peripheral <strong>and</strong> centralnervous system has already been d<strong>on</strong>e in [113] with no improvement of the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance.However the features computed from the EEG signals were <strong>on</strong>ly st<strong>and</strong>ard features (see secti<strong>on</strong>3.3.1) not specially related to emoti<strong>on</strong>al processes which could explain their low per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance <str<strong>on</strong>g>for</str<strong>on</strong>g>emoti<strong>on</strong> assessment.In this study c<strong>on</strong>catenati<strong>on</strong> of the features was d<strong>on</strong>e to combine the features extracted from thedifferent peripheral sensors (GSR, Respirati<strong>on</strong> belt, etc.) in a unique feature set. It was also usedto fuse peripheral features with features computed from the EEG signals. C<strong>on</strong>catenating N Ffeature sets F 1 , …,jfi<str<strong>on</strong>g>for</str<strong>on</strong>g> each sample i <strong>and</strong> all feature set j:NFF in a new feature set F c<strong>on</strong>sists in the c<strong>on</strong>catenati<strong>on</strong> of feature vectorsf f f f(4.13)1 j N[ Fii i i]4.3.2 Classifier levelAccording to S<strong>and</strong>ers<strong>on</strong> et al. [153] the fusi<strong>on</strong> at the classifier level (or post-mapping fusi<strong>on</strong>) isdivided in two categories: decisi<strong>on</strong> fusi<strong>on</strong> <strong>and</strong> opini<strong>on</strong> fusi<strong>on</strong>. In decisi<strong>on</strong> fusi<strong>on</strong>, the estimatedclasses given by several classifiers are combined to decide which class will be attributed to asample. The most comm<strong>on</strong> method <str<strong>on</strong>g>for</str<strong>on</strong>g> decisi<strong>on</strong> fusi<strong>on</strong> is certainly majority voting: the class thathas been chosen by most of the classifiers is attributed to the sample. However, decisi<strong>on</strong> fusi<strong>on</strong>strategies are generally applicable <strong>on</strong>ly in particular cases (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance the number of classifier isc<strong>on</strong>strained by the number of classes <str<strong>on</strong>g>for</str<strong>on</strong>g> majority voting). This is why the present study focuses<strong>on</strong> opini<strong>on</strong> fusi<strong>on</strong>.a. Opini<strong>on</strong> fusi<strong>on</strong>: sum ruleIn opini<strong>on</strong> fusi<strong>on</strong>, a classifier outputs a score <str<strong>on</strong>g>for</str<strong>on</strong>g> each class. This score generally represents theclassifier c<strong>on</strong>fidence that the test sample bel<strong>on</strong>gs to a given class. Product <strong>and</strong> sum rules can thenbe applied to fuse the scores given by multiple classifiers [153]. In this work, the probabilisticoutputs of the classifiers were used as a measure of c<strong>on</strong>fidence. The following sum rule wasapplied to fuse those output probabilities <str<strong>on</strong>g>for</str<strong>on</strong>g> a class i:85


Chapter 4kiqQCj1qQP( | f )qiP ( | f )qjqQ1QP ( | f )qi(4.14)where Q is the ensemble of the classifiers chosen <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong>, |Q| the number of such classifiers<strong>and</strong> P q ( i | f) is the posterior probability of having class i according to classifier q. The finalchoice is d<strong>on</strong>e by selecting the class i with the highest value k i . It can be observed that k i canalso be viewed as a c<strong>on</strong>fidence measure <strong>on</strong> the class given by the fusi<strong>on</strong> of classifiers.b. Opini<strong>on</strong> fusi<strong>on</strong>: Bayes belief integrati<strong>on</strong>One drawback of the sum rule is that it does not take into account the errors made by each of theclassifiers [154]. As an extreme example, it is possible that a (very bad) classifier q classifiesmost of the instances of a class i into a class j . Thus, when the class estimate y ˆqof thisclassifier is i there is a high probability that the true class be in fact j . For Bayes beliefintegrati<strong>on</strong> [154], the errors produced by the classifiers are expressed by the probabilitiesP( | yˆ) computed from the c<strong>on</strong>fusi<strong>on</strong> matrices obtained from the training set. The fusi<strong>on</strong> isiqthen per<str<strong>on</strong>g>for</str<strong>on</strong>g>med by assuming classifiers independency <strong>and</strong> choosing the class i that maximizesthe following probability:P( | yˆ)i qqQˆ ˆi 1 | Q| | Q| 1P( i)P( | y ... y )(4.15)where Q is the ensemble of classifiers used <str<strong>on</strong>g>for</str<strong>on</strong>g> the fusi<strong>on</strong> <strong>and</strong> y ˆqis the class estimate of classifierq. Notice that the probability P( ˆ ˆi| y1 ... y| Q|) could be computed directly from the training setwithout any assumpti<strong>on</strong> <strong>on</strong> classifiers independence. However it is generally difficult to estimateit reliably because of the low number of training samples compared to the high number ofcombinati<strong>on</strong>s of the |Q| classifiers estimates.4.4 Rejecti<strong>on</strong> of samplesIn classificati<strong>on</strong>, the samples that lie close to the decisi<strong>on</strong> boundary can be c<strong>on</strong>sidered as lessreliably classified than those that are far from the decisi<strong>on</strong> boundary. For this reas<strong>on</strong> a classifiercan assign a class to a sample <strong>on</strong>ly if it is sufficiently far away from the decisi<strong>on</strong> boundary (i.e.the c<strong>on</strong>fidence in the classificati<strong>on</strong> is sufficiently high).As a final step, rejecti<strong>on</strong> of trials that have a c<strong>on</strong>fidence value k i (determined using the methodpresented in Secti<strong>on</strong> 4.3.2.a) below a threshold reject was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med to improve classificati<strong>on</strong>86


Methods <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessmentaccuracy. When a sample is rejected because the c<strong>on</strong>fidence value is not sufficiently high, thesample is not classified <strong>and</strong> the classificati<strong>on</strong> accuracy is computed <strong>on</strong>ly <strong>on</strong> the samples with highc<strong>on</strong>fidence. The percentage of rejected samples as well as the accuracy computed <strong>on</strong> theremaining samples thus become a functi<strong>on</strong> of the reject threshold. A good value <str<strong>on</strong>g>for</str<strong>on</strong>g> reject wouldbe the <strong>on</strong>e which provides a compromise between accuracy maximizati<strong>on</strong> <strong>and</strong> rejecti<strong>on</strong> rateminimizati<strong>on</strong>. This approach is taken is chapter 6.87


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s elicited by visual stimuliChapter 5<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s elicited by visual stimuli5.1 Introducti<strong>on</strong>Am<strong>on</strong>g all the senses, the visual modality is the <strong>on</strong>e that has the largest capacity (i.e. it can carrya large amount of in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> in a given time) [155]. It is thus not surprising that, in HMI, mostof the in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> is communicated to the user through the visual modality (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance viascreens). C<strong>on</strong>sequently, recognizing visually elicited emoti<strong>on</strong>s is important to enhance affectivecomputing methods. In this chapter, images were chosen as visual stimuli to elicit emoti<strong>on</strong>s. Theper<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of several methods to assess the valence <strong>and</strong> arousal dimensi<strong>on</strong>s of these emoti<strong>on</strong>sfrom physiological signals was then analyzed. Since this work uses images as stimuli, its resultscan also have some impact <strong>on</strong> the automatic affective annotati<strong>on</strong> of multimedia data (see Secti<strong>on</strong>1.2.3.c).Secti<strong>on</strong> 5.2 describes how an emoti<strong>on</strong>al database of physiological features was c<strong>on</strong>structed. InSecti<strong>on</strong> 5.3 different classes were defined in the valence-arousal space <strong>and</strong> the classificati<strong>on</strong>methodology employed to recognize them are presented. Secti<strong>on</strong> 5.4 discusses the resultsobtained <strong>and</strong> stresses the interest of EEG features al<strong>on</strong>e as well as fused with other peripheralfeatures in emoti<strong>on</strong> assessment.5.2 Data collecti<strong>on</strong>This secti<strong>on</strong> details the creati<strong>on</strong> of a database of physiological features where emoti<strong>on</strong>s, definedas classes in the valence-arousal space, where elicited by using images. It explains how the imagestimuli where chosen, describes the protocol designed to acquire the data <strong>and</strong> finally presents theextracted features.5.2.1 Visual stimuliIn this study we used a subset of images from the 700 emoti<strong>on</strong>ally evocative pictures of the IAPS(Internati<strong>on</strong>al <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Picture System [54]). Each of these images has been extensivelyevaluated by north-American participants, providing valence <strong>and</strong> arousal values <strong>on</strong> nine pointsscales (ranging from 1 to 9). Experimentati<strong>on</strong> showed a 0.8 correlati<strong>on</strong> with evaluati<strong>on</strong>sper<str<strong>on</strong>g>for</str<strong>on</strong>g>med by Europeans [156]. Each image is associated with average arousal µ A <strong>and</strong> valence µ Vcomputed from these evaluati<strong>on</strong>s as well as with their st<strong>and</strong>ard deviati<strong>on</strong>s A <strong>and</strong> V . However, asobserved during our experiments, feelings induced by an image <strong>on</strong> a particular participant can bevery different from the <strong>on</strong>es expected. This is likely due to difference in past experience. Selfassessmentof valence/arousal was there<str<strong>on</strong>g>for</str<strong>on</strong>g>e per<str<strong>on</strong>g>for</str<strong>on</strong>g>med in the present study by each participant<strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> each image. Two different subsets of the IAPS images were presented to the participants,<strong>on</strong>e to study the valence dimensi<strong>on</strong> (valence subset) <strong>and</strong> the other to study the arousal dimensi<strong>on</strong>89


Chapter 5(arousal subset) of emoti<strong>on</strong>s. The complete list of the pictures presented to the participants can befound in Appendix C.To c<strong>on</strong>struct the valence subset, 50 negative <strong>and</strong> 50 positive images were selected according tothe following c<strong>on</strong>straints:negative images: 5; 3; 2 ;A V Vpositive images: 5; 7; 2.A V VThe c<strong>on</strong>straint <strong>on</strong> the mean arousal allows keeping <strong>on</strong>ly images that have high arousal <strong>and</strong> thusintense positive or negative c<strong>on</strong>tent. The c<strong>on</strong>straint <strong>on</strong> the mean valence is useful to distinguishnegative from positive images while a low st<strong>and</strong>ard deviati<strong>on</strong> ensures stability in the judgmentsof the participants (i.e. there is a high probability that a participant rates an image with a valenceclose to the mean IAPS valence). Pictures were than r<strong>and</strong>omly selected from the subset of theIAPS images satisfying the c<strong>on</strong>straints.For the arousal subset, 50 images of high arousal <strong>and</strong> 50 images of low arousal were selectedThe pictures were taken from the 3 subsets of the IAPS images defined by the followingc<strong>on</strong>straints:low arousal <strong>and</strong> neutral: 3; 4 6;high arousal <strong>and</strong> positive: 5.5; 5;high arousal <strong>and</strong> negative: 5.5; 5.50 images were chosen r<strong>and</strong>omly from the “low arousal <strong>and</strong> neutral” set to c<strong>on</strong>struct the final lowarousal set. To c<strong>on</strong>struct the high arousal set 25 images were chosen from both the “high arousal<strong>and</strong> positive” <strong>and</strong> the “high arousal <strong>and</strong> negative” sets. This procedure was employed becausethere are few pictures with high arousal <strong>and</strong> neutral valence. It is important that the high arousalset c<strong>on</strong>tains the same number of negative <strong>and</strong> positive images to ensure that the arousal axis willbe assessed at the classificati<strong>on</strong> stage (<strong>and</strong> not the difference between neutral <strong>and</strong> negativeimages as could be the case if <strong>on</strong>ly negative image were present in this set). No c<strong>on</strong>straint wasadded <strong>on</strong> the variance with the disadvantages that participants’ evaluati<strong>on</strong>s can str<strong>on</strong>gly differfrom the IAPS evaluati<strong>on</strong>s. However, if a c<strong>on</strong>straint <strong>on</strong> the st<strong>and</strong>ard deviati<strong>on</strong> would be added,the number of images would have been too low to c<strong>on</strong>struct sets of 50 pictures.By using the c<strong>on</strong>straints above, some of the high arousal images of the arousal subset can also bepart of the valence subset. This should be avoided because the arousal <strong>and</strong> valence sets wereAAAVVV90


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s elicited by visual stimulipresented c<strong>on</strong>secutively to the participants (see Secti<strong>on</strong> 5.2.2) <strong>and</strong> the sec<strong>on</strong>d time an image ispresented, the intensity of the felt emoti<strong>on</strong> can be weakened. For this reas<strong>on</strong> the images werechosen in order to minimize the intersecti<strong>on</strong> of the arousal <strong>and</strong> valence subsets. This result in atotal of 19 images present in both sets (the list of these images is given in Appendix C).5.2.2 Acquisiti<strong>on</strong> protocolWe acquired data from 4 participants, 3 males, 1 female, aged from 28 to 49. One of theparticipants is left h<strong>and</strong>ed. Cortical activity was acquired by recording an EEG with 64 electrodes(see Secti<strong>on</strong> 3.1.1). The peripheral sensors used were the GSR sensor, the plethysmograph tomeasure BVP, the respirati<strong>on</strong> belt to evaluate abdominal <strong>and</strong> thoracic movements, <strong>and</strong> thetemperature sensor. All signals were sampled at 1024 Hz.For each experimental recording, the participant equipped with the above sensors was sitting in fr<strong>on</strong>tof a computer screen in a bare room relatively immune to electromagnetic noise. The valence <strong>and</strong> thearousal sets of images were presented c<strong>on</strong>secutively to the participant. For both sets, each trialcorresp<strong>on</strong>ding to a visual stimulus was scheduled as described in Figure 5.1. A dark screen was firstdisplayed <str<strong>on</strong>g>for</str<strong>on</strong>g> 3 sec<strong>on</strong>ds to “rest <strong>and</strong> prepare” the participant <str<strong>on</strong>g>for</str<strong>on</strong>g> the next image. A white cross wasthen drawn <strong>on</strong> the screen center <str<strong>on</strong>g>for</str<strong>on</strong>g> a r<strong>and</strong>om period of 2 to 4 sec<strong>on</strong>ds, to attract user's attenti<strong>on</strong> <strong>and</strong>avoid accustoming. An IAPS image was subsequently displayed <str<strong>on</strong>g>for</str<strong>on</strong>g> 6 sec<strong>on</strong>ds, while at the same timea trigger was sent <str<strong>on</strong>g>for</str<strong>on</strong>g> synchr<strong>on</strong>izati<strong>on</strong>. The task of the participant was to watch each image <strong>and</strong> selfassess the valence <strong>and</strong> the arousal of his / her emoti<strong>on</strong> using a simplified versi<strong>on</strong> of the SAM [73](see Figure 5.1). The valence/arousal scales were composed of 5 possible numerical judgments <str<strong>on</strong>g>for</str<strong>on</strong>g>each dimensi<strong>on</strong> with scales ranging from -2 to 2 <str<strong>on</strong>g>for</str<strong>on</strong>g> valence <strong>and</strong> 0 to 4 <str<strong>on</strong>g>for</str<strong>on</strong>g> arousal. This selfassessmentstep was not limited in time to allow <str<strong>on</strong>g>for</str<strong>on</strong>g> a resting period between images.-2 2Valence experimentArousal experimentTrial 1 Trial 2Trial 100timetime0 4Dark screen White cross Display image3s2-4s6sSAM?stimeFigure 5.1. Descripti<strong>on</strong> of the acquisiti<strong>on</strong> protocol. (left) the modified SAM used <str<strong>on</strong>g>for</str<strong>on</strong>g> self assessment. (right) theschedule of the protocol.According to the 5 factors defined by Picard [106] this protocol corresp<strong>on</strong>ds to an openrecordingc<strong>on</strong>diti<strong>on</strong> since the participant knew that his / her physiological activity was recorded.91


Chapter 5The emoti<strong>on</strong>s were event-elicited using the images. The emphasis was put <strong>on</strong> the feeling ofemoti<strong>on</strong>s rather than the expressi<strong>on</strong> of the emoti<strong>on</strong> since participants had to self-assess theirfeelings. The complete recording was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med in a lab setting (special room dedicated to therecording).5.2.1 Features extractedFor the EEG signals two feature sets were used <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> (see Secti<strong>on</strong> 3.3.2.a):- the EEG_Lateral feature set was employed <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> <strong>on</strong> the valence dimensi<strong>on</strong>;- the EEG_Area feature set was used <str<strong>on</strong>g>for</str<strong>on</strong>g> the classificati<strong>on</strong> <strong>on</strong> the arousal dimensi<strong>on</strong>.The choice of the EEG_Lateral feature set is motivated by the fact that studies have shown thatthe asymmetry index can be useful to discriminate positive from negative stimuli while Aftanaset al. [82] dem<strong>on</strong>strated that the EEG_Area features are significantly different <str<strong>on</strong>g>for</str<strong>on</strong>g> low <strong>and</strong> higharousal stimuli. Notice that most of the EEG_Area features c<strong>on</strong>cern the Occipital (O) lobe, which isinteresting since this lobe corresp<strong>on</strong>ds to the visual cortex <strong>and</strong> subjects are stimulated with pictures.C<strong>on</strong>cerning peripheral signals, HR was estimated from the blood pressure signal by using themethod presented in Secti<strong>on</strong> 3.2.4. The 5 peripheral signals to analyze are there<str<strong>on</strong>g>for</str<strong>on</strong>g>e: GSR, BP,HR, respirati<strong>on</strong> <strong>and</strong> temperature. Table 5.1 presents the features extracted from each of thesesignals over the 6 sec<strong>on</strong>ds epoch. They corresp<strong>on</strong>d to some of the st<strong>and</strong>ard features presented inSecti<strong>on</strong> 3.3.1. A total of 18 features were thus obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> the peripheral signals. Since no postprocessingalgorithm was applied to improve the peak detecti<strong>on</strong> <strong>on</strong> short time BVP signals (seeSecti<strong>on</strong> 3.2.4) <strong>and</strong> since the minimum <strong>and</strong> maximum features are sensible to outliers, thosefeatures were not computed <str<strong>on</strong>g>for</str<strong>on</strong>g> the HR signal.Signal xxMinxMaxxBVP X X X XHR X XGSR X X X XRespirati<strong>on</strong> X X X XTemperature X X X XTable 5.1. The 18 features extracted from the peripheral signals.For arousal classificati<strong>on</strong>, the EEG features <strong>and</strong> the peripheral features were c<strong>on</strong>catenated aspresented in secti<strong>on</strong> 4.3.1 to analyze the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of fusi<strong>on</strong> of peripheral <strong>and</strong> centralin<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> at the feature level. This step was not taken <str<strong>on</strong>g>for</str<strong>on</strong>g> valence classificati<strong>on</strong> as <str<strong>on</strong>g>for</str<strong>on</strong>g> thisproblem the classificati<strong>on</strong> accuracy obtained with peripheral features was at the r<strong>and</strong>om level (seeSecti<strong>on</strong> 5.4.1).92


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s elicited by visual stimuli5.3 Classificati<strong>on</strong>5.3.1 Ground-truth definiti<strong>on</strong>sThe visual stimuli used in both the valence <strong>and</strong> arousal experiments were purposely chosen tobel<strong>on</strong>g to two distinct classes: negative vs. positive <str<strong>on</strong>g>for</str<strong>on</strong>g> the valence experiment <strong>and</strong> calm vs.excited <str<strong>on</strong>g>for</str<strong>on</strong>g> the arousal experiment. Since self-evaluati<strong>on</strong>s were also collected, the ground-truthcan be defined either a-priori, based <strong>on</strong> the classes defined by the IAPS evaluati<strong>on</strong>s, or a-posteriori using the self-evaluati<strong>on</strong>s. Some of the advantages <strong>and</strong> disadvantages of these twomethods are discussed in Secti<strong>on</strong> 4.1.1.This secti<strong>on</strong> compares the IAPS evaluati<strong>on</strong>s to the self-assessment values <strong>and</strong> discusses thec<strong>on</strong>structi<strong>on</strong> of the ground-truth <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment. For this purpose, the IAPS evaluati<strong>on</strong>s(ranging from 1 to 9) were linearly projected in the same range as the self-assessment values(ranging from -2 to 2 <str<strong>on</strong>g>for</str<strong>on</strong>g> valence <strong>and</strong> 0 to 4 <str<strong>on</strong>g>for</str<strong>on</strong>g> arousal) using the following <str<strong>on</strong>g>for</str<strong>on</strong>g>mulas:( AIAPS1)A 2( VIAPS 5)V 2(5.1)V IAPS <strong>and</strong> A IAPS being the original IAPS valence / arousal values <strong>and</strong> V <strong>and</strong> A being the newvalence / arousal values.a. Valence experimentAs can be seen from Figure 5.2 the valence distributi<strong>on</strong> of the self-evaluati<strong>on</strong>s is very close to the<strong>on</strong>e obtained from the IAPS evaluati<strong>on</strong>s. This tends to validate that the visual stimuli elicited theexpected emoti<strong>on</strong>s: either positive or negative emoti<strong>on</strong>s. However, 3 of the 4 participants judgedsome of the stimuli as being of neutral valence with 79% of those stimuli bel<strong>on</strong>ging to thepositive class according to the IAPS evaluati<strong>on</strong>s. Moreover <strong>on</strong>ly <strong>on</strong>e participant ranked three ofthe stimuli with a value of 2. Those results dem<strong>on</strong>strate that, according to the 4 participants selfassessments,positive emoti<strong>on</strong>s were more difficult to elicit than negative emoti<strong>on</strong>s.C<strong>on</strong>cerning arousal evaluati<strong>on</strong>s, the IAPS <strong>and</strong> self-assessment distributi<strong>on</strong>s were found to bequite different. For instance most of the stimuli were self-evaluated with an arousal value of 1while, because of the c<strong>on</strong>straints applied <str<strong>on</strong>g>for</str<strong>on</strong>g> the selecti<strong>on</strong> of images, n<strong>on</strong>e of the stimuli had anIAPS arousal value below 2. This result can be explained by two factors. First, the stimuli mayhave elicited lower arousal than expected. Sec<strong>on</strong>dly, the participants may have used the completescale to report <str<strong>on</strong>g>for</str<strong>on</strong>g> the arousal difference between the stimuli (which is a measure relative to theset of stimuli) rather than to report absolute arousal value. When looking at the IAPS arousal93


Chapter 5histogram with a higher number of bins than in Figure 5.2 (which is possible because the IAPSvalues are means computed from a 9 points scale), the histogram computed in the interval [2,4] isthen very similar to the <strong>on</strong>e obtained from the self-assessment. This encourages the argument ofrelative self-assessment <strong>and</strong> shows that the self-assessed arousal values were not so differentfrom the IAPS evaluati<strong>on</strong>s.Since the self-assessments were close to the IAPS evaluati<strong>on</strong>s, particularly <str<strong>on</strong>g>for</str<strong>on</strong>g> valenceevaluati<strong>on</strong>s, the ground-truth was defined based <strong>on</strong> the IAPS evaluati<strong>on</strong>s <strong>on</strong>ly. This allows toc<strong>on</strong>struct two classes of interest: <strong>on</strong>e class corresp<strong>on</strong>ding to the positive stimuli <strong>and</strong> <strong>on</strong>ecorresp<strong>on</strong>ding to the negative stimuli.Original IAPS evaluati<strong>on</strong>sSelf-assessmentsNumber of IAPS imagesNumber of IAPS imagesValenceValenceNumber of IAPS imagesNumber of IAPS imagesArousalArousalFigure 5.2. Histograms of the IAPS <strong>and</strong> self evaluati<strong>on</strong>s (valence <strong>and</strong> arousal) <str<strong>on</strong>g>for</str<strong>on</strong>g> the valence experiment. Foreasier comparis<strong>on</strong> of IAPS evaluati<strong>on</strong>s <strong>and</strong> self evaluati<strong>on</strong>s the IAPS values have been normalized to the samerange as the self evaluati<strong>on</strong>s.b. Arousal experimentAs <str<strong>on</strong>g>for</str<strong>on</strong>g> the valence experiment, the distributi<strong>on</strong> of valence obtained from the IAPS <strong>and</strong> selfassessmentvalues are quite similar (see Figure 5.3). The higher number of images self-evaluatedas having a valence of 1 compared to the IAPS evaluati<strong>on</strong>s is mostly due to participants94


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s elicited by visual stimulievaluating originally neutral images (valence value to 0) as slightly positive. This is particularlytrue <str<strong>on</strong>g>for</str<strong>on</strong>g> participant 3 as can be seen from Figure 5.3.Original IAPS evaluati<strong>on</strong>sSelf-assessmentsNumber of IAPS imagesNumber of IAPS imagesValenceValenceNumber of IAPS imagesNumber of IAPS imagesArousalArousalFigure 5.3. Histograms of the IAPS <strong>and</strong> self evaluati<strong>on</strong>s (valence <strong>and</strong> arousal) <str<strong>on</strong>g>for</str<strong>on</strong>g> the arousal experiment. Foreasier comparis<strong>on</strong> of IAPS evaluati<strong>on</strong>s <strong>and</strong> self evaluati<strong>on</strong>s the IAPS values have been normalized in the samerange as the self evaluati<strong>on</strong>s.For arousal, the histograms from the IAPS <strong>and</strong> self-assessment values were again different. Dueto the c<strong>on</strong>straints applied to c<strong>on</strong>struct sets of low <strong>and</strong> high arousal stimuli, two peaks can beobserved <str<strong>on</strong>g>for</str<strong>on</strong>g> the arousal IAPS histogram. However, those peaks are not clearly visible <str<strong>on</strong>g>for</str<strong>on</strong>g> thehistogram obtained from self-assessments. Only participant 3 obtained similar peaks with a highnumber of images rated with arousal values of 0 <strong>and</strong> 2. Since no c<strong>on</strong>straint was applied <strong>on</strong> thevariance of arousal during the selecti<strong>on</strong> of the stimuli, the histogram difference is certainly due toa large variability of the arousal judgments across participants <strong>and</strong> dem<strong>on</strong>strates the difference inevaluati<strong>on</strong> that can be observed <str<strong>on</strong>g>for</str<strong>on</strong>g> the same stimuli.Since the distributi<strong>on</strong>s of self-evaluati<strong>on</strong>s <strong>and</strong> IAPS values were found to be different <str<strong>on</strong>g>for</str<strong>on</strong>g> arousal<strong>and</strong> the purpose of this experiment is to assess the arousal dimensi<strong>on</strong> of emoti<strong>on</strong>s, different setsof classes were c<strong>on</strong>structed based <strong>on</strong> either the IAPS values or the self-assessments.95


Chapter 5The images used <str<strong>on</strong>g>for</str<strong>on</strong>g> the arousal assessment were purposely chosen to be of either very low orvery high IAPS arousal values, that is they essentially should have bel<strong>on</strong>ged to 2 classes. For thisreas<strong>on</strong>, when using the IAPS judgments as a basis to build ground-truth classes, it was natural todivide data into two sets, <strong>on</strong>e <str<strong>on</strong>g>for</str<strong>on</strong>g> the calm emoti<strong>on</strong>s <strong>and</strong> the other <str<strong>on</strong>g>for</str<strong>on</strong>g> the exciting emoti<strong>on</strong>s. Inthis way, two well balanced ground-truth classes of 50 patterns each were obtained.It is more difficult to determine classes from the self-assessment values. As shown by thehistograms of arousal, the evaluati<strong>on</strong>s are not equally distributed across the 5 choices <strong>and</strong> inparticular do not readily corresp<strong>on</strong>d to 2 classes. Taking this into account, two differentclassificati<strong>on</strong> experiments based <strong>on</strong> the self-assessment were d<strong>on</strong>e:96- with 2 ground-truth classes, were the calm class c<strong>on</strong>tained patterns judged in the calmestcategory <strong>and</strong> the exiting class the others,- with 3 ground-truth classes (calm, neutral, exciting) were the calm class corresp<strong>on</strong>ded tothe first of the 5 judgment values, the neutral class to the sec<strong>on</strong>d <strong>and</strong> third, <strong>and</strong> theexciting class to the last two.Both class definiti<strong>on</strong>s led to unbalanced classes, especially <str<strong>on</strong>g>for</str<strong>on</strong>g> the 3-classes problem: the excitingclass c<strong>on</strong>tained very few samples (6 to 23 depending <strong>on</strong> the participant) compared to the calmclass (32 to 45) <strong>and</strong> the neutral class (32 to 55).5.3.2 MethodsAn ANOVA test was applied <strong>on</strong> the features of the EEG_Lateral feature set to c<strong>on</strong>trol thatsignificant differences were observed in the asymmetry scores between the positively <strong>and</strong>negatively valenced emoti<strong>on</strong>al states. This was d<strong>on</strong>e to verify the precedent findings c<strong>on</strong>cerningalpha lateralizati<strong>on</strong> in the case where emoti<strong>on</strong>s are stimulated by pictures <strong>and</strong> also to check if theasymmetry scores can be used as features <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of classificati<strong>on</strong>. Since classificati<strong>on</strong>was d<strong>on</strong>e in an intra-participant framework (a model was designed <str<strong>on</strong>g>for</str<strong>on</strong>g> each participant) theANOVA test was run separately <str<strong>on</strong>g>for</str<strong>on</strong>g> each participant. For the EEG_Area feature set, no ANOVAtest was applied because Aftanas et al. [82] already dem<strong>on</strong>strated the interest of those features <str<strong>on</strong>g>for</str<strong>on</strong>g>arousal discriminati<strong>on</strong> in a protocol very similar to the <strong>on</strong>e proposed in this chapter.A Naïve Bayes classifier was first applied <strong>on</strong> the features of each participant. This classifier isknown to be optimal under the assumpti<strong>on</strong> of c<strong>on</strong>diti<strong>on</strong>ally independent features <strong>and</strong> in the caseof complete knowledge of the underlying probability distributi<strong>on</strong>s of the problem. Modeling theunderlying distributi<strong>on</strong>s is un<str<strong>on</strong>g>for</str<strong>on</strong>g>tunately difficult in our study, since very few samples areavailable to c<strong>on</strong>struct them; a per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance decrease is thus unavoidable. For the sake ofcomparis<strong>on</strong>, classificati<strong>on</strong> based <strong>on</strong> LDA was also per<str<strong>on</strong>g>for</str<strong>on</strong>g>med. In this case the distributi<strong>on</strong>s areassumed to be multivariate Gaussians with no assumpti<strong>on</strong> of independence. Due to the rather


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s elicited by visual stimulilimited number of patterns, a leave <strong>on</strong>e out cross validati<strong>on</strong> was preferred to a k-fold strategy inorder to maximize the size of the training set (see Secti<strong>on</strong> 4.1.2). Results presented in the nextsecti<strong>on</strong> are the percentage of well classified examples.5.4 Results5.4.1 Valence experimentAn ANOVA test was applied <strong>on</strong> the EEG_Lateral features to check <str<strong>on</strong>g>for</str<strong>on</strong>g> a difference in meanbetween the two valence c<strong>on</strong>diti<strong>on</strong>s (Table 5.2). C<strong>on</strong>trary to what was expected from Davids<strong>on</strong>’stheory [19, 81], no significant differences were found in the fr<strong>on</strong>tal area. This could be due to thea lower signal to noise ratio in this regi<strong>on</strong> because of facial muscular artifacts that were notremoved. Another explanati<strong>on</strong> could be that the lateralizati<strong>on</strong> of alpha waves was dem<strong>on</strong>strated<str<strong>on</strong>g>for</str<strong>on</strong>g> approach <strong>and</strong> withdrawal stimuli [19] which are different from positive <strong>and</strong> negative stimuli.However the lateralizati<strong>on</strong> was significant in areas located more at the rear of the brain such asparietal <strong>and</strong> occipital areas. Those results could partly be explained by the nature of the stimulisince visual processing takes place in occipital areas. Moreover, the absence of alpha hemisphericlateralizati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the fr<strong>on</strong>tal areas <strong>and</strong> a significant effect <str<strong>on</strong>g>for</str<strong>on</strong>g> the parietal area (P3-P4 electrodes)was also found in [157] where the participants had to mentally review film sequences. Sincesome of the features were shown to have significantly different means between the twoc<strong>on</strong>diti<strong>on</strong>s, this feature set was used <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of classificati<strong>on</strong>.Electrodes pairsp-valuesPaticipant 1 Paticipant 2 Paticipant 3 Paticipant 4Fp2-Fp1 0.78 0.62 0.57 0.30AF4-AF3 0.22 0.71 0.59 0.70F4-F3 0.72 0.43 0.27 0.79FC4-FC3 0.83 0.17 0.80 0.06C4-C3 0.52 0.20 0.53 0.45CP4-CP3 0.75 0.08 0.68 0.04P4-P3 0.39 0.10 0.09 0.13PO4-PO3 0.01 0.08 0.01 0.12O2-O1 0.11 < 0.01 0.59 0.01Table 5.2. p-values of the ANOVA test applied <strong>on</strong> the lateralizati<strong>on</strong> features <str<strong>on</strong>g>for</str<strong>on</strong>g> the two groups defined by theIAPS classes (negative vs. positive visual stimuli) <strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> each participant. p-values < 0.1 are highlited in gray.The Naïve Bayes classifier accuracies obtained <strong>on</strong> both the peripheral <strong>and</strong> EEG features were allaround the r<strong>and</strong>om level of 50%. Since better results were obtained with an LDA (Figure 5.4) <strong>on</strong>the EEG feature set this suggest that the Naïve Bayes assumpti<strong>on</strong> of c<strong>on</strong>diti<strong>on</strong>al independence isnot valid in this case. By looking at the covariance matrices of the normalized EEG features <str<strong>on</strong>g>for</str<strong>on</strong>g>the two classes, some features were found to be correlated. This shows the EEG featuresc<strong>on</strong>diti<strong>on</strong>al dependency <strong>and</strong> explains the poor results of the Naïve Bayes classifier.97


Chapter 5The accuracies obtained with the LDA <str<strong>on</strong>g>for</str<strong>on</strong>g> the peripheral <strong>and</strong> EEG features are presented inFigure 5.4. The mean bars <strong>on</strong> the right of Figure 5.4 represent the average accuracy obtained <str<strong>on</strong>g>for</str<strong>on</strong>g>the 4 participants (P1 to P4). As can be seen from this figure, the accuracy is higher than ther<strong>and</strong>om level <str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG features <strong>and</strong> around the r<strong>and</strong>om level <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features. Theaverage EEG accuracy across participants is of 58%. This result dem<strong>on</strong>strates the importance ofEEG features <str<strong>on</strong>g>for</str<strong>on</strong>g> better assessment of the valence dimensi<strong>on</strong> of emoti<strong>on</strong>s <strong>and</strong> shows that EEGsignals should not be neglected <str<strong>on</strong>g>for</str<strong>on</strong>g> the assessment of emoti<strong>on</strong>s from physiological signals. Thelower accuracy obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features is not surprising since the peripheral activity,more specifically the aut<strong>on</strong>omous nervous system activity, is known to better correlate with thearousal dimensi<strong>on</strong> than with the valence dimensi<strong>on</strong> of emoti<strong>on</strong>s [7, 87].70%60%50%Accuracy40%30%20%10%EEGPeripheral0%P1 P2 P3 P4 MeanFigure 5.4. LDA accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> of negative <strong>and</strong> positive stimuli.985.4.2 Arousal experimentWhen using the two arousal ground-truth classes defined according to the IAPS judgment, theNaïve Bayes classifier average accuracy across participants exceeded the chance level <strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g>EEG features (54% vs. 50%). The LDA classifier per<str<strong>on</strong>g>for</str<strong>on</strong>g>med slightly better, with an averageaccuracy of 55%, 53% <strong>and</strong> 54% <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG, physiological <strong>and</strong> fused features respectively. Thoserelatively low accuracies are likely due to large differences between the IAPS values <strong>and</strong> theactual emoti<strong>on</strong> felt by the participant (as detailed in secti<strong>on</strong> 5.3.1.b). We c<strong>on</strong>cluded that in ourexperimental setting the IAPS arousal judgments could not be recovered from actualphysiological measurements, <strong>and</strong> had to use self-assessments. However, using the LDA, theaccuracy of peripheral features is higher than the r<strong>and</strong>om level. This c<strong>on</strong>firms that the peripheralfeatures computed in this study are more suitable <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> of the arousal dimensi<strong>on</strong> thanthe valence dimensi<strong>on</strong> of emoti<strong>on</strong>s.Results with ground-truth classes obtained from self-evaluati<strong>on</strong>s are presented in Figure 5.5 <strong>and</strong>Figure 5.6. The percentage of well classified patterns <str<strong>on</strong>g>for</str<strong>on</strong>g> the four participants <strong>and</strong> the averageacross participants are shown. Compared to accuracies obtained with the ground-truth defined by


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s elicited by visual stimulithe IAPS judgments, accuracies obtained with the self-assessed ground-truth are higher,especially <str<strong>on</strong>g>for</str<strong>on</strong>g> participants 2 <strong>and</strong> 3 (Figure 5.5). This tends to c<strong>on</strong>firm that physiological signalsbetter correlate with pers<strong>on</strong>alized self assessment of emoti<strong>on</strong> than with the generalized IAPSjudgments. The best per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of 72% is obtained by using the EEG signals of participant 2<strong>and</strong> a Naïve Bayes classifier. A similar result is obtained with the LDA (70%). For bothclassifiers, the average accuracy obtained with EEG signals is higher than the <strong>on</strong>e obtained fromperipheral signals. Those results stress again the added value of using EEG signals <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>alassessment.Accuracy80%70%60%50%40%30%20%10%0%Naive BayesP1 P2 P3 P4 Mean80%70%60%50%40%30%20%10%0%P1 P2 P3 P4 MeanFigure 5.5. Classifiers accuracy with 2 classes c<strong>on</strong>structed from self-assessment.LDAEEGPhysioC<strong>on</strong>catenati<strong>on</strong>It is worth menti<strong>on</strong>ing that there are two drawbacks to the problem of unbalanced classes in thecase three arousal classes are assessed: (i) some of the classes are str<strong>on</strong>gly undersampled <strong>and</strong> (ii)the accuracy measure is less reliable since there are more samples bel<strong>on</strong>ging to <strong>on</strong>e of the classthan to the others. The first drawback implies that the probability distributi<strong>on</strong>s of theundersampled classes cannot be correctly determined which results in a weak assessment of thoseclasses. C<strong>on</strong>cerning the sec<strong>on</strong>d drawback, since the number of samples in each class was equal<str<strong>on</strong>g>for</str<strong>on</strong>g> the two feature sets <strong>and</strong> similar across participants, we believe that the comparis<strong>on</strong> of theemoti<strong>on</strong> assessment per<str<strong>on</strong>g>for</str<strong>on</strong>g>mances based <strong>on</strong> this measure of accuracy is still reliable.Figure 5.6 shows results <str<strong>on</strong>g>for</str<strong>on</strong>g> the three class problem. Again, the features extracted from the EEGof participant 2 yield the best result of 58% of well classified patterns (compared to a chancelevel of 33%). Participant 4 still obtained the worst accuracy. This is likely due to the highnumber of eye-blinks that were found in the EEG signals of this participant (approximately <strong>on</strong>eblink per sec<strong>on</strong>d). Participant 1 obtained better results with a Bayes classifier than with a LDA.Extreme results <str<strong>on</strong>g>for</str<strong>on</strong>g> participants 2 can be explained by a better underst<strong>and</strong>ing of the selfassessment procedure since he had a good knowledge about emoti<strong>on</strong>s, <strong>and</strong> was likely toaccurately evaluate his feelings.The results obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong> by c<strong>on</strong>catenati<strong>on</strong> are different depending <strong>on</strong> the participant, theclassifier <strong>and</strong> the number of defined classes. For the Naïve-Bayes classifier, the c<strong>on</strong>catenati<strong>on</strong> ofperipheral <strong>and</strong> EEG features slightly increased the average accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> the 3 arousal classes <strong>and</strong>decreased it <str<strong>on</strong>g>for</str<strong>on</strong>g> 2 arousal classes. For the LDA, c<strong>on</strong>catenati<strong>on</strong> of features increased the average99


Chapter 5accuracy by 5% <str<strong>on</strong>g>for</str<strong>on</strong>g> 3 arousal classes <strong>and</strong> did not affect it in the other case. Thus the LDA seemsmore appropriate <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong> at the feature level, which could be explained by the weakassumpti<strong>on</strong> of c<strong>on</strong>diti<strong>on</strong>al independence of the Naïve-Bayes classifier. Also, fusi<strong>on</strong> providesmore robust results since some participants had better scores with peripheral signals than withEEG's <strong>and</strong> vice-versa.Accuracy60%50%40%30%20%10%0%Naive BayesLDA60%50%40%30%20%10%0%P1 P2 P3 P4 MeanP1 P2 P3 P4 MeanFigure 5.6. Classifiers accuracy with 3 classes c<strong>on</strong>structed from self-assessment.EEGPhysioC<strong>on</strong>catenati<strong>on</strong>5.5 C<strong>on</strong>clusi<strong>on</strong>In this chapter two categories of physiological signals, from the central <strong>and</strong> from the peripheralnervous systems, have been evaluated <strong>on</strong> the problem of assessing the arousal <strong>and</strong> the valencedimensi<strong>on</strong> of emoti<strong>on</strong>s elicited by IAPS images. Those assessments were per<str<strong>on</strong>g>for</str<strong>on</strong>g>med asclassificati<strong>on</strong> problems, with ground-truth valence / arousal values provided either by the IAPS orby self-assessments of the emoti<strong>on</strong>. Two classifiers were used, a Naïve-Bayes classifier <strong>and</strong> aLDA.Results showed the usability of EEG's in both arousal <strong>and</strong> valence recogniti<strong>on</strong> <strong>and</strong> the interest ofEEG features over peripheral features. Moreover, the fusi<strong>on</strong> of EEG features with peripheralfeatures improved the assessment per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance. This improvement was better with a LDA thanwith the Naïve-Bayes classifier. Results also markedly improved when using classes generatedfrom self-assessment of emoti<strong>on</strong>s. When trying to assess emoti<strong>on</strong>s, <strong>on</strong>e should avoid usingpredefined labels but rather ask <str<strong>on</strong>g>for</str<strong>on</strong>g> the user’s feeling. However, by using self-assessment thegenerated classes were unbalanced which gave rise to classificati<strong>on</strong> problems such as theundersampling of some classes. Moreover, using the self-assessments as a ground-truth impliesthat the user is an expert in evaluating his / her feelings, which is not always the case as discussedin Secti<strong>on</strong> 4.1.1.Future work <strong>on</strong> arousal assessment will first aim at improving <strong>on</strong> the current results by using othern<strong>on</strong>-linear classifiers, such as Support Vector Machines. Feature selecti<strong>on</strong> <strong>and</strong> more sophisticatedfusi<strong>on</strong> strategies will also be examined, jointly with the examinati<strong>on</strong> of other features such astemporal characteristics of signals that are known to be str<strong>on</strong>gly implied in emoti<strong>on</strong>al processes.100


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>sChapter 6<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>s6.1 Introducti<strong>on</strong>Fairly recent psychological studies regarding the relati<strong>on</strong>s between emoti<strong>on</strong>s <strong>and</strong> the brain haveuncovered the str<strong>on</strong>g involvement of cognitive processes in emoti<strong>on</strong>s [61, 79-81, 84, 158].Am<strong>on</strong>g those studies, some used the recall of past emoti<strong>on</strong>al events to self-induce emoti<strong>on</strong>s [84,158]. This method has the advantage of activating many brain areas because cognitive processesrelated to memory retrieval are located throughout the brain. Asking participants to self-induceemoti<strong>on</strong>s is also useful because the remembered emoti<strong>on</strong>s will corresp<strong>on</strong>d with the emoti<strong>on</strong> thatis required by the protocol. This allows <str<strong>on</strong>g>for</str<strong>on</strong>g> an optimal c<strong>on</strong>trol over the number of emoti<strong>on</strong>s thatare elicited per class. From a classificati<strong>on</strong> point of view, this is interesting because it avoids theproblem of unbalanced classes encountered in Chapter 5. Other advantages of this elicitati<strong>on</strong>method are presented in Secti<strong>on</strong> 4.1.1. To our knowledge, despite of all those advantages, thismethod has never been used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment from EEG features (see Secti<strong>on</strong> 2.3).Having in mind the previous c<strong>on</strong>siderati<strong>on</strong>s, the present Chapter aims at investigating theusefulness of EEG <strong>and</strong> peripheral signals in a self-inducti<strong>on</strong> paradigm. In order to be as muchapplicati<strong>on</strong> independent as possible, we used the valence-arousal space as a prior model to definethree emoti<strong>on</strong>al classes of interest that are calm-neutral, positive-excited <strong>and</strong> negative-excited.The protocol <strong>and</strong> the features computed from the recorded signals are presented in Secti<strong>on</strong> 6.2.Secti<strong>on</strong> 6.3 describes the complete framework used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> recogniti<strong>on</strong>. Finally, results arepresented discussed in Secti<strong>on</strong> 6.4.6.2 Data acquisiti<strong>on</strong>6.2.1 Acquisiti<strong>on</strong> protocolIn [106] five factors that can influence recordings were defined: subject-elicited vs. eventelicited,laboratory setting vs. real world, focus <strong>on</strong> expressi<strong>on</strong> vs. feeling of the emoti<strong>on</strong>, openlyrecordedvs. hidden recording <strong>and</strong> emoti<strong>on</strong>-purpose vs. other-purpose. The following protocoldescripti<strong>on</strong> addresses those five factors as well as those emphasized in Secti<strong>on</strong> 2.3.In the present study a self-induced method, using recall of str<strong>on</strong>g emoti<strong>on</strong>al episodes, isemployed to elicit reliable <strong>and</strong> short time emoti<strong>on</strong>s. An episode is defined as a situati<strong>on</strong> thatlasted <str<strong>on</strong>g>for</str<strong>on</strong>g> a several minutes <strong>and</strong> potentially c<strong>on</strong>taining several events <strong>and</strong> acti<strong>on</strong>s with the sameemoti<strong>on</strong>al orientati<strong>on</strong>. An example is the funeral of a relative including events such as momentsof the cerem<strong>on</strong>y <strong>and</strong> the burial in itself. The elicited emoti<strong>on</strong>s are c<strong>on</strong>sidered reliable because (i)thinking of the same episodes ought to produce similar reacti<strong>on</strong>s from <strong>on</strong>e trial to another, (ii)emoti<strong>on</strong>al episodes are often stored in memory because the emoti<strong>on</strong>s felt were quite intense,(iii)101


Chapter 6emoti<strong>on</strong>al recall is a cognitive task that induces EEG activity [84, 158] as well as modifyperipheral activity [6, 92].Compared to other studies [93, 106, 111], where emoti<strong>on</strong>s are elicited <strong>and</strong> assessed over severalminutes, the durati<strong>on</strong> of an emoti<strong>on</strong> epoch in this study is merely 8 s. This epoch durati<strong>on</strong> waschosen because it is the maximum durati<strong>on</strong> that allows maintaining the total length of theprotocol below <strong>on</strong>e hour to avoid participant fatigue. Within the requirement of the <strong>on</strong>e hourdurati<strong>on</strong> this epoch is maximized <str<strong>on</strong>g>for</str<strong>on</strong>g> three reas<strong>on</strong>s. Firstly, some peripheral features need to bedetermined over a sufficiently l<strong>on</strong>g period of time in order to be reliably computed; this is <str<strong>on</strong>g>for</str<strong>on</strong>g>instance the case <str<strong>on</strong>g>for</str<strong>on</strong>g> statistical features extracted from HR. In general, an epoch of 8 s shouldsuffice <str<strong>on</strong>g>for</str<strong>on</strong>g> this purpose if we do not c<strong>on</strong>sider the very low frequency features such as lowfrequency HRV [96]. Sec<strong>on</strong>dly, recalling past episodes <strong>and</strong> eliciting the corresp<strong>on</strong>ding emoti<strong>on</strong>sare difficult tasks <strong>and</strong> participants might need a few sec<strong>on</strong>ds to accomplish them. Thirdly, thereacti<strong>on</strong> time of peripheral activity from the moment where the emoti<strong>on</strong> is elicited is of severalsec<strong>on</strong>ds, with the GSR being the slowest resp<strong>on</strong>se with a lag around 1-4 sec<strong>on</strong>ds [88].The 11 participants (7 males, 4 females) who took part in the study were aged from 26 to 40, <strong>on</strong>ebeing left-h<strong>and</strong>ed. One week be<str<strong>on</strong>g>for</str<strong>on</strong>g>e the recording, participants were told to retrieve from theirmemory <strong>on</strong>e excited-positive <strong>and</strong> <strong>on</strong>e excited-negative episode that had occurred in their life <strong>and</strong>that they c<strong>on</strong>sider as being most powerful. On the day of the experiment, each participant wasgiven a c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m where the c<strong>on</strong>text, the goal <strong>and</strong> a short explanati<strong>on</strong> of the experiment wereprovided. Participants had to sign this c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m to c<strong>on</strong>tinue further <strong>and</strong> could stop theexperiment whenever they wanted. After signing the c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m, sensors were attached to theparticipant who was seated in fr<strong>on</strong>t of a computer screen. A precise descripti<strong>on</strong> of the protocolwas provided with a support dem<strong>on</strong>strati<strong>on</strong>. This type of experiment corresp<strong>on</strong>ds to Picard’sfactors <str<strong>on</strong>g>for</str<strong>on</strong>g> an open-recording (participants knew they were recorded), emoti<strong>on</strong>-purpose(participants new the objective of the study), <strong>and</strong> laboratory settings (participant are recorded in ac<strong>on</strong>trolled envir<strong>on</strong>ment).The complete recording sessi<strong>on</strong> was divided into trials. During each trial participants had toaccomplish a particular task according to the visual cue displayed <strong>on</strong> the m<strong>on</strong>itor after a r<strong>and</strong>omdurati<strong>on</strong> display of a dark screen (Figure 6.1). This task could be to self-generate <strong>on</strong>e of the twoexcited emoti<strong>on</strong>s by using the past emoti<strong>on</strong>al episodes of their life as a support, or to stay calm<strong>and</strong> relax in order to define a third emoti<strong>on</strong>al state called calm-neutral. A total of T = 300 trials(100 trials per emoti<strong>on</strong>al state) were per<str<strong>on</strong>g>for</str<strong>on</strong>g>med in a r<strong>and</strong>om order. Since facial muscle artifactscan c<strong>on</strong>taminate EEG signals, participants were encouraged not to express their feelings throughfacial expressi<strong>on</strong>s, not to blink, <strong>and</strong> not to close their eyes during the 8 s of recordings (despite ofthis instructi<strong>on</strong> some involuntary facial expressi<strong>on</strong>s artifacts can still occur in the signals).Emphasis was thus put <strong>on</strong> the feeling of emoti<strong>on</strong>s rather than <strong>on</strong> the cognitive task of102


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>sremembering <strong>and</strong> <strong>on</strong> the motor expressi<strong>on</strong>s of emoti<strong>on</strong>s. A resting period of unlimited durati<strong>on</strong> torelax <strong>and</strong> stretch muscles was proposed to participants after each block of 30 trials. As can beseen from Figure 6.1, the chosen emoti<strong>on</strong>al states do not cover all areas of the valence-arousalspace, especially in the bottom half of the space. This choice was made because there are actuallyfew emoti<strong>on</strong>s that are calm-negative or calm-positive [23, 54].ExcitedTrial 1 Trial 30 Rest Trial 31 Trial 300NegativePositivetimeCalmDark screen2-3sDisplay image corresp<strong>on</strong>ding to the desiredemoti<strong>on</strong>8stimeFigure 6.1. (left) The different emoti<strong>on</strong>al classes represented in the valence-arousal space <strong>and</strong> their associatedimage. (right) schedule of the protocol <strong>and</strong> detail of a trial.Data were recorded using the Biosemi Active II system. EEG signals were recorded using 64surface electrodes. Plugged <strong>on</strong> the same system to simplify synchr<strong>on</strong>izati<strong>on</strong>, other sensors wereused to record peripheral activity: a GSR (Galvanic Skin Resp<strong>on</strong>se) sensor to evaluate sudati<strong>on</strong>, arespirati<strong>on</strong> belt to record abdominal expansi<strong>on</strong> <strong>and</strong> a plethysmograph to measure blood pressure.Both EEG <strong>and</strong> peripheral signals were sampled at 1024 Hz.After data acquisiti<strong>on</strong>, participants were asked to report verbally <strong>on</strong> their experiences in anin<str<strong>on</strong>g>for</str<strong>on</strong>g>mal interview. Participants were not asked to provide a detailed descripti<strong>on</strong> of the chosenepisodes because we believe that <str<strong>on</strong>g>for</str<strong>on</strong>g> pers<strong>on</strong>al <strong>and</strong> ethical reas<strong>on</strong>s a participant may hesitate torefer to his / her str<strong>on</strong>gest emoti<strong>on</strong>al experiences. For this reas<strong>on</strong> the differences in the cognitivetasks between different trials could not be fully c<strong>on</strong>trolled. However, as argued in [84] the knowneffectiveness of mental imagery as an elicitator of powerful emoti<strong>on</strong>s can compensate thisproblem.The present protocol <str<strong>on</strong>g>for</str<strong>on</strong>g> off-line acquisiti<strong>on</strong> of physiological signals is very close to thoseencountered in the BCI community so that the c<strong>on</strong>clusi<strong>on</strong>s drawn from this study may also havesome impact in this directi<strong>on</strong>. An emoti<strong>on</strong> elicitati<strong>on</strong> task can then be regarded as a mental taskthat the user tries to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m in order to communicate his / her feelings. This can be useful <str<strong>on</strong>g>for</str<strong>on</strong>g>severely disabled people that cannot directly express their emoti<strong>on</strong>s. Current BCI paradigms [27,29, 30] aim to detect brain activity that corresp<strong>on</strong>ds to complex tasks (mental calculus,imaginati<strong>on</strong> of finger taping, etc.) not related to the objective of the user (moving a mouse cursor,choosing a letter, etc.). Generally the user needs training be<str<strong>on</strong>g>for</str<strong>on</strong>g>e using such systems. In case the103


Chapter 6objective of the user is to express an emoti<strong>on</strong>, classical BCI tasks (e.g., imaginati<strong>on</strong> of fingertapping) seem to be really far from this objective <strong>and</strong> it is more appropriate to use tasks such asthe remembering of a similar emoti<strong>on</strong>al episode.6.2.2 Feature extracti<strong>on</strong>From the EEG signals, 2 feature sets were extracted to analyze their per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>assessment. Those features were computed <strong>on</strong> the last 7.5 sec<strong>on</strong>ds of the signal. The first 0.5sec<strong>on</strong>ds were removed <str<strong>on</strong>g>for</str<strong>on</strong>g> the two following reas<strong>on</strong>s. Firstly, this time window was not expectedto c<strong>on</strong>tain emoti<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> related to the requested emoti<strong>on</strong>al episode since it is ratherunlikely that a participant start to remember the episode directly after the display of theassociated image. Sec<strong>on</strong>dly, this part of the signal c<strong>on</strong>tains the P300 [159] wave <strong>and</strong> may c<strong>on</strong>tainemoti<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> related to the nature of the presented image (smiling or sad smiley) whichis not the stimulus studied here.The 2 feature sets computed from the EEG signals are (See Secti<strong>on</strong> 3.3.2.a):- the EEG_STFT feature set, obtained from the spectrogram of the EEG signals, c<strong>on</strong>tainingin that case 96429 16704 features (9 frequency b<strong>and</strong>s, 64 electrodes <strong>and</strong> 29 timeframes);- the EEG_MI feature set, c<strong>on</strong>taining the mutual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> between each pairs of the 64electrodes with a total of (64 1)64 2016 features.2As can be seen, both feature sets are of very high dimensi<strong>on</strong>ality.Peripheral signal St<strong>and</strong>ard features Advanced features ReferencexxxGSR X X DecRate DecTimefGSR, fGSRSecti<strong>on</strong> 3.3.2.bBVPXHeart Rate (HR) X X XRespirati<strong>on</strong> X X XPowfResp,DRf Secti<strong>on</strong> 3.3.2.eRespTable 6.1. The features extracted from the peripheral signalsThe peripheral features extracted from the corresp<strong>on</strong>ding signals are given in Table 6.1. Adetailed explanati<strong>on</strong> of the features can be found in Secti<strong>on</strong> 3.3.1 <str<strong>on</strong>g>for</str<strong>on</strong>g> the st<strong>and</strong>ard features. The“Reference” column indicates the secti<strong>on</strong> of Chapter 3 that corresp<strong>on</strong>ds to an explanati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> thegiven advanced features. All the peripheral features were computed from the complete durati<strong>on</strong>of the 8 sec<strong>on</strong>ds trial <strong>and</strong> c<strong>on</strong>catenated in a feature vector. The HR signal was computed from theBVP signal of the plethysmograph as discussed in Secti<strong>on</strong> 3.2.4. No a-posteriori correcti<strong>on</strong> of the104


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>swr<strong>on</strong>gly detected beats was applied because of the short durati<strong>on</strong> of a trial. For a given trial, allthe peripheral features were c<strong>on</strong>catenated in a unique feature vector c<strong>on</strong>taining a total of 18features. The peripheral feature set c<strong>on</strong>structed by that way was called Peripheral.6.3 Classificati<strong>on</strong>6.3.1 The different classificati<strong>on</strong> schemesFrom the protocol detailed in Secti<strong>on</strong> 6.2.1, the ground-thruth is easily defined by attributing toeach trial the corresp<strong>on</strong>ding label of calm, positive-excited <strong>and</strong> negative-excited. However, toanalyze the classificati<strong>on</strong> per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance in different classes’ <str<strong>on</strong>g>for</str<strong>on</strong>g>mulati<strong>on</strong>s, the five classificati<strong>on</strong>tasks described bellow were tested.EEG <strong>and</strong>peripheralsignalsTrial 1PreprocessingType of classificati<strong>on</strong>task (CPN, PN, CP,CN, CE)Trial / classes selecti<strong>on</strong> <strong>and</strong> grouping :Type of stimuliTrial 2…Trial 300<strong>and</strong>featureextracti<strong>on</strong> CPN : 300 trials, y i { c, p, n} NP : 200 trials, y i { p, n} CP : 200 trials, y i { c, p} CN : 200 trials, y i { c, n} CE : 300 trials, y i { c, e}, e = n U pQ classifiers(<strong>on</strong>e per feature set)Trial 1Trial 2…Trial 3003 feature sets <strong>and</strong> the associatedlabels (300 trials) : peripheral (18 features) ; EEG_STFT (16704 features) ;Classificati<strong>on</strong>(with or without featureselecti<strong>on</strong>)Leave <strong>on</strong>e out cross validati<strong>on</strong>Accuracy ofambiguity rejecti<strong>on</strong>Accept classificati<strong>on</strong>of this trialYesk i > rejectNok iP( | f )qFusi<strong>on</strong> of the Qclassifiers by thesummati<strong>on</strong> ruleiAccuracy of eachclassifier qAccuracy of fusi<strong>on</strong>Reject this trialFigure 6.2. Complete process of trial acquisiti<strong>on</strong>, classificati<strong>on</strong>, fusi<strong>on</strong> <strong>and</strong> rejecti<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> a given participant. Asdefined in Chapter 4, k i is the c<strong>on</strong>fidence measure of class i after opini<strong>on</strong> fusi<strong>on</strong> <strong>and</strong> reject is the rejecti<strong>on</strong>threshold.Since each recorded trial corresp<strong>on</strong>ds to a particular emoti<strong>on</strong>al state, it is easy to <str<strong>on</strong>g>for</str<strong>on</strong>g>mulate aclassificati<strong>on</strong> task (called CPN <str<strong>on</strong>g>for</str<strong>on</strong>g> "calm", "positive", "negative") where the three ground-truthclasses c , p , n corresp<strong>on</strong>d to calm-neutral, positive-excited <strong>and</strong> negative-excited patterns. Atarget class vector y CPN = [y 1 …,y i ,…,y 300 ] T is c<strong>on</strong>structed, where y i { c , p , n } represents the105


Chapter 6class of the trial i. We also address other classificati<strong>on</strong> tasks by c<strong>on</strong>structing different targetvectors to distinguish between the following emoti<strong>on</strong>al states: negative excited vs. positiveexcited(NP), calm-neutral vs. positive-excited (CP), calm-neutral vs. negative excited (CN),calm vs. excited (CE) by regrouping samples of the positive-excited <strong>and</strong> negative-excited states.As summarized in Figure 6.2, there are three sets of features, Peripheral , EEG_STFT <strong>and</strong>EEG_MI that c<strong>on</strong>tain respectively peripheral features, STFT EEG features <strong>and</strong> MI EEG features<str<strong>on</strong>g>for</str<strong>on</strong>g> all trials. Those feature sets are associated with the class vectors y CPN , y NP , y CP , y CN <strong>and</strong> y CE ,depending <strong>on</strong> the classificati<strong>on</strong> task to address.6.3.2 ClassifiersSince the EEG feature sets (EEG_STFT <strong>and</strong> EEG_MI) are of very high dimensi<strong>on</strong>ality (thous<strong>and</strong>sof features) compared to the number of samples in the sets (200 or 300 depending <strong>on</strong> theclassificati<strong>on</strong> scheme), there is always a linear boundary that can completely separate trainingsamples of the different classes. For this reas<strong>on</strong> <strong>on</strong>ly linear classifiers were applied <strong>on</strong> thosefeature sets. Another advantage of using linear classifiers is that they give better generalizedsoluti<strong>on</strong>s. The issue of feature space reducti<strong>on</strong> was also investigated as detailed in Secti<strong>on</strong> 6.3.3.The following linear classifiers were applied <strong>on</strong> the EEG feature sets:- the LDA because it can provide probabilistic output which is useful <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose offusi<strong>on</strong>. Since the EEG feature spaces are of high dimensi<strong>on</strong>ally it sometimes occurred thatthe covariance matrix of the features was singular, in that case the diag<strong>on</strong>alized versi<strong>on</strong> ofthe LDA was employed;- the linear SVM since they are known to have good per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance in high dimensi<strong>on</strong>alspaces [138];- the probabilistic linear SVM to obtain probabilistic output;- the RVM since, as the SVM, it should per<str<strong>on</strong>g>for</str<strong>on</strong>g>m well in high dimensi<strong>on</strong>al spaces <strong>and</strong>provides probabilistic outputs.The above menti<strong>on</strong>ed classifiers were also applied <strong>on</strong> the Peripheral feature set. Since thisfeature set is of a lower dimensi<strong>on</strong>ality than the EEG feature sets, the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mances of thefollowing n<strong>on</strong> linear classifiers were also investigated:- the QDA to obtain quadratic decisi<strong>on</strong> boundaries together with probabilistic outputs;- the RBF SVM, where the RBF kernel size was chosen according to the procedureexplained in Secti<strong>on</strong> 4.1.3.c;106


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>s- the probabilistic RBF SVM to obtain probabilistic output.In the present study, different classifiers were trained <strong>on</strong> the three feature sets to recover theground truth classes. A classifier was trained separately <str<strong>on</strong>g>for</str<strong>on</strong>g> each participant <strong>and</strong> the accuracy ofeach classifier was evaluated using the leave-<strong>on</strong>e-out strategy. This involves using each featurevector in turn as the test set <strong>and</strong> the remaining <strong>on</strong>es as the learning set. At each step, a classifier istrained from the learning set <strong>and</strong> then applied to the test sample. This leave-<strong>on</strong>e-out strategy waschosen since it provides the maximum possible size of the learning set. This is preferable in thisproblem because the number of samples (N=300) is very low compared to the size of the EEGfeature spaces.6.3.3 Reducti<strong>on</strong> of the feature spaceIn order to study the effects of feature space reducti<strong>on</strong> <strong>on</strong> the classificati<strong>on</strong> accuracy, severalmethods were tested: the Fast Correlati<strong>on</strong> <str<strong>on</strong>g>Based</str<strong>on</strong>g> Filter (FCBF) presented in Secti<strong>on</strong> 4.2.1c <strong>and</strong>filtering based <strong>on</strong> the Fisher criteri<strong>on</strong> presented in Secti<strong>on</strong> 4.2.1b. The classificati<strong>on</strong> accuracy ofthe LDA <strong>and</strong> the linear SVM were evaluated <str<strong>on</strong>g>for</str<strong>on</strong>g> different parameter values of these featureselecti<strong>on</strong> methods. Since both these methods are computati<strong>on</strong>ally expensive, they were tested <strong>on</strong><strong>on</strong>e participant <strong>on</strong>ly, namely participant 1 (who was the first recorded participant). The reducti<strong>on</strong>of the EEG_STFT feature set was addressed since this is the <strong>on</strong>e that c<strong>on</strong>tains the higher numberof features.The FCBF parameter of the FCBF algorithm represents the minimum correlati<strong>on</strong> value with theclass labels y that a feature f should have in order to be selected. This parameter is thus related tothe number of selected features <strong>and</strong> the higher the parameter the lower the number of selectedfeatures. In this study, the value of this parameter ranged from 0.05 to 0.3 with a step of 0.05. Thehigher bound of 0.3 was chosen because most of the features had a correlati<strong>on</strong> value below thisvalue (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance, <str<strong>on</strong>g>for</str<strong>on</strong>g> the CPN classificati<strong>on</strong> scheme <strong>on</strong>ly <strong>on</strong>e feature has a correlati<strong>on</strong> valuehigher than 0.3). At the lower bound of 0.05 approximately half of the features are removed inthe first step of the FCBF algorithm.The Fisher criteri<strong>on</strong> defined in Secti<strong>on</strong> 4.2.1b can be used to rank the features by relevance order.It is then possible to select <strong>on</strong>ly the most relevant features <strong>and</strong> filter out the others, which allowsto compute the classificati<strong>on</strong> accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> a given number of selected features. In this study, theper<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of this filter technique was studied <str<strong>on</strong>g>for</str<strong>on</strong>g> numbers of features taken in the interval [1F EEG_STFT ] (where F EEG_STFT is the number of features in the complete EEG_STFT feature set).This method was used <strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g> the CPN classificati<strong>on</strong> scheme.107


Chapter 66.3.4 Fusi<strong>on</strong> <strong>and</strong> rejecti<strong>on</strong>To investigate the advantages of fusi<strong>on</strong> of several physiological feature sets, classificati<strong>on</strong>accuracy was studied <str<strong>on</strong>g>for</str<strong>on</strong>g> both of the methods presented in Secti<strong>on</strong> 4.3: fusi<strong>on</strong> at the feature level<strong>and</strong> fusi<strong>on</strong> at the classifier level. Fusi<strong>on</strong> was d<strong>on</strong>e independently <str<strong>on</strong>g>for</str<strong>on</strong>g> each participant <strong>and</strong> thecombinati<strong>on</strong> of the following feature sets was studied:- EEG_STFT <strong>and</strong> EEG_MI feature sets, to analyze the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of fusi<strong>on</strong> of differentfeature sets computed from the same EEG modality,- EEG_STFT, EEG_MI <strong>and</strong> Peripheral to analyze the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of fusi<strong>on</strong> of in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>originating from the CNS with in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> originating from the PNS.Fusi<strong>on</strong> at the feature level was d<strong>on</strong>e by the c<strong>on</strong>catenati<strong>on</strong> of features as explained in Secti<strong>on</strong>4.3.1. The fusi<strong>on</strong> at the classifier level was d<strong>on</strong>e by opini<strong>on</strong> fusi<strong>on</strong>, using the sum rule detailed inSecti<strong>on</strong> 4.3.2a. In this case, a classifier was chosen <str<strong>on</strong>g>for</str<strong>on</strong>g> each feature set to <str<strong>on</strong>g>for</str<strong>on</strong>g>m the ensemble Q ofclassifiers (with |Q| = 2 or |Q| = 3 depending <strong>on</strong> the number of feature sets used <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong>). Theensemble Q was composed of the classifiers with probabilistic outputs that generally had the bestaccuracy <strong>on</strong> the associated feature set.As a final step the method detailed in Secti<strong>on</strong> 4.4 to reject samples with a low c<strong>on</strong>fidence k i valuewas applied to the CPN, NP <strong>and</strong> CE classificati<strong>on</strong> scheme. Those schemes were chosen becausethey are the most relevant <str<strong>on</strong>g>for</str<strong>on</strong>g> HCI applicati<strong>on</strong>s. To analyze the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of this rejecti<strong>on</strong>, theaverage accuracy across participants computed <strong>on</strong> the n<strong>on</strong>-rejected samples <strong>and</strong> the percentage ofrejected samples were plotted as a functi<strong>on</strong> of the reject threshold (taken in the range [0 1]). Sincethe label of each trial is already determined after fusi<strong>on</strong>, it is possible to compare the number ofbadly classified trials that are rejected to the correctly classified <strong>on</strong>es. The sums over participantsof the correctly classified <strong>and</strong> badly classified samples that are rejected are also plotted as afuncti<strong>on</strong> of the reject threshold. From the analysis of those curves, a value <str<strong>on</strong>g>for</str<strong>on</strong>g> the reject thresholdthat improves the accuracy while keeping the percentage of samples reas<strong>on</strong>ably low will besuggested.6.4 Results6.4.1 Participants reports <strong>and</strong> protocol validati<strong>on</strong>Out of the 11 recorded participants 10 reported a successful elicitati<strong>on</strong> of the emoti<strong>on</strong>s byrecalling emoti<strong>on</strong>al episodes. As can be seen from Figure 6.5, which represents the averageaccuracies obtained from the 10 participants cited above, the peripheral activity is useful todistinguish between different classes of emoti<strong>on</strong>s. This implies that different patterns ofphysiological activity where induced <str<strong>on</strong>g>for</str<strong>on</strong>g> each emoti<strong>on</strong>al task <strong>and</strong> thus supports the idea that108


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>semoti<strong>on</strong>s were successfully elicited. However, all participants reported that it was really difficultto stay c<strong>on</strong>centrated throughout the entire recording. A recurring observati<strong>on</strong> was also thatswitching from <strong>on</strong>e emoti<strong>on</strong> to another very quickly was sometimes c<strong>on</strong>fusing <strong>and</strong> hard toaccomplish. The effects of such observati<strong>on</strong>s can be missing trials where the participants did notaccomplish the requested task, the elicitati<strong>on</strong> of the undesired boredom emoti<strong>on</strong> which caninterfere with positive <strong>and</strong> negative excited trials, <strong>and</strong> noisy EEG signals due to fatigue. Theemoti<strong>on</strong>s used in the two excited categories were also different from a participant to another (<str<strong>on</strong>g>for</str<strong>on</strong>g>instance a participant elicited a negative-excited emoti<strong>on</strong> by remembering a past episode wherehe felt fear while another participant used an episode where he felt anger).In the protocol presented in Secti<strong>on</strong> 6.2.1 brain activity can be induced by two cognitivecomp<strong>on</strong>ents: the actual events of the episode (<str<strong>on</strong>g>for</str<strong>on</strong>g> instance thinking of some<strong>on</strong>e crying) <strong>and</strong> theemoti<strong>on</strong> elicitati<strong>on</strong> following the event. Since our aim is to detect emoti<strong>on</strong>s, it is important toc<strong>on</strong>trol that the events used to induce emoti<strong>on</strong>s were not always the same to ensure that what isdetected from brain signals is the emoti<strong>on</strong> <strong>and</strong> not the cognitive task related to the event (<str<strong>on</strong>g>for</str<strong>on</strong>g>instance mental imagery of the act of crying). Since participants did not report about the episodesthey used to induce emoti<strong>on</strong>s it is difficult to c<strong>on</strong>trol <str<strong>on</strong>g>for</str<strong>on</strong>g> this, however the following remarks leadus to assume the protocol is valid:- two participants reported that they thought of different episodes within the same category(i.e. positive-excited <strong>and</strong> negative-excited). The classificati<strong>on</strong> accuracy obtained from thesignals of <strong>on</strong>e of those two participants actually corresp<strong>on</strong>ds to the best results across the 10participants while the other <strong>on</strong>e obtained average accuracies;- <strong>on</strong>e participant reported that he thought to the episodes without c<strong>on</strong>centrating <strong>on</strong> the feelingof emoti<strong>on</strong>s which resulted in a weak emoti<strong>on</strong> elicitati<strong>on</strong>. All the accuracies computed fromthe signals of this participant are at the r<strong>and</strong>om level;- since an episode was defined as including several emoti<strong>on</strong>al events of the same category, it isunlikely that the participants always thought of the same event to elicit <strong>on</strong>e of the emoti<strong>on</strong>s;- the participants were explicitly told to focus <strong>on</strong> the feeling of emoti<strong>on</strong>s <strong>and</strong> emoti<strong>on</strong>s weresuccessfully elicited as stated above.Notice that the participant who did not c<strong>on</strong>centrate <strong>on</strong> the feeling of emoti<strong>on</strong>s was removed <str<strong>on</strong>g>for</str<strong>on</strong>g>further analysis since he did not follow the protocol properly.6.4.2 Results of single classifiersFigure 6.3, Figure 6.4 <strong>and</strong> Figure 6.5 respectively present the mean accuracy across participants<str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG_STFT features, the EEG_MI features <strong>and</strong> the Peripheral features. The accuracies ofdifferent modalities <strong>and</strong> classifiers are compared below to answer the following questi<strong>on</strong>s: what109


Chapter 6is the effectiveness of EEG <strong>and</strong> peripheral features to assess emoti<strong>on</strong>s according to the differentclassificati<strong>on</strong> schemes <strong>and</strong> which classifiers should be used <str<strong>on</strong>g>for</str<strong>on</strong>g> latter fusi<strong>on</strong> of feature sets.The EEG_STFT features provided interesting results with a mean classificati<strong>on</strong> accuracy of 63%<str<strong>on</strong>g>for</str<strong>on</strong>g> three classes <strong>and</strong> a SVM classifier (the r<strong>and</strong>om level is at 33% accuracy). The best averageaccuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> two classes is obtained from the CP classificati<strong>on</strong> task with nearly 80% of wellclassified trials (r<strong>and</strong>om level at 50%), followed by the CE <strong>and</strong> NP classificati<strong>on</strong> tasks withrespectively 78% <strong>and</strong> 74% of accuracy. For all participants <strong>and</strong> all classificati<strong>on</strong> tasks, the resultsare higher than the r<strong>and</strong>om levels (33% <str<strong>on</strong>g>for</str<strong>on</strong>g> three classes <strong>and</strong> 50% <str<strong>on</strong>g>for</str<strong>on</strong>g> two classes). The EEG_MIfeatures seem to be a bit less suitable <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> classificati<strong>on</strong> than EEG_STFT features with anapproximate decrease of well classified trials of 2% to 4%, except <str<strong>on</strong>g>for</str<strong>on</strong>g> the NP classificati<strong>on</strong> taskwhere a slight per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance increase was noted. It is hard to compare those results to the state ofthe art because there are <strong>on</strong>ly few studies using EEG. In the previous study reported in Chapter 5the best accuracy <strong>on</strong> two <strong>and</strong> three arousal classes was respectively of 72% <strong>and</strong> 58%. In thisstudy the highest accuracies <str<strong>on</strong>g>for</str<strong>on</strong>g> the CE <strong>and</strong> CPN classificati<strong>on</strong> tasks are respectively of 88% <strong>and</strong>86.3%. The best result <str<strong>on</strong>g>for</str<strong>on</strong>g> a two class task is obtained <strong>on</strong> the NP task with 96% of accuracy. In[115] an accuracy of 29% was obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> 6 different emoti<strong>on</strong>al classes while the accuracy wasof 42% <str<strong>on</strong>g>for</str<strong>on</strong>g> identificati<strong>on</strong> of 5 emoti<strong>on</strong>al states in [113]. Our results are thus superior to theprevious studies using the EEG modality <str<strong>on</strong>g>for</str<strong>on</strong>g> detecting emoti<strong>on</strong>s expressed in the valence-arousalspace <strong>and</strong> in alignment with results obtained <strong>on</strong> emoti<strong>on</strong>al labels.100%90%80%70%Accuracy60%50%40%30%20%10%LDALinear SVMProb. Linear SVMRVM0%CPN CE NP CN CPFigure 6.3. Mean classifier accuracy across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG_STFT feature set <strong>and</strong> the differentclassificati<strong>on</strong> schemes. The bars <strong>on</strong> top of each column represents the st<strong>and</strong>ard deviati<strong>on</strong> across participants.110


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>sAccuracy100%90%80%70%60%50%40%30%20%10%LDALinear SVMProb. Linear SVMRVM0%CPN CE NP CN CPFigure 6.4. Mean classifier accuracy across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG_MI feature set <strong>and</strong> the differentclassificati<strong>on</strong> schemes. The bars <strong>on</strong> top of each column represents the st<strong>and</strong>ard deviati<strong>on</strong> across participants.To check <str<strong>on</strong>g>for</str<strong>on</strong>g> the usability of this emoti<strong>on</strong>al protocol as a new BCI paradigm our results werecompared to BCI accuracies. In [160] the authors showed that around 75% of the 99 untrainedparticipants that took part in a two class BCI paradigm without feedback obtained accuraciesbetween 60% <strong>and</strong> 79%. The distributi<strong>on</strong>s of the accuracies <str<strong>on</strong>g>for</str<strong>on</strong>g> our recall paradigm are similar;however more participants should be recorded to validate this statement. Our results are also farfrom those of more recent BCI studies where the accuracy can reach more than 90% <str<strong>on</strong>g>for</str<strong>on</strong>g> twoclasses <str<strong>on</strong>g>for</str<strong>on</strong>g> many untrained participants [30]. This can be due to the definiti<strong>on</strong> of mental task thatare chosen to activate well separated areas of the brain, c<strong>on</strong>trary to the task definiti<strong>on</strong> used in thisstudy.100%90%80%Accuracy70%60%50%40%30%20%10%LDAQDALinear SVMProb. linear SVMRBF SVMProb. RBF SVMLinear RVM0%CPN CE NP CN CPFigure 6.5. Mean classifier accuracy across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features <strong>and</strong> the different classificati<strong>on</strong>schemes. The bars <strong>on</strong> top of each column represents the st<strong>and</strong>ard deviati<strong>on</strong> across participants.111


Chapter 6If the three figures (Figure 6.3, Figure 6.4 <strong>and</strong> Figure 6.5) are compared, it is obvious that theEEG features lead to better accuracy than peripheral features <str<strong>on</strong>g>for</str<strong>on</strong>g> all classificati<strong>on</strong> schemes. Forperipheral features the LDA classifier is the best with an average accuracy of 51% <str<strong>on</strong>g>for</str<strong>on</strong>g> threeclasses <strong>and</strong> around 66% <str<strong>on</strong>g>for</str<strong>on</strong>g> two classes (except <str<strong>on</strong>g>for</str<strong>on</strong>g> the NP classificati<strong>on</strong> task with accuracyaround 61%). Results ranged from nearly the r<strong>and</strong>om level up to around 80% <str<strong>on</strong>g>for</str<strong>on</strong>g> two class<str<strong>on</strong>g>for</str<strong>on</strong>g>mulati<strong>on</strong>s <strong>and</strong> from 37% up to 75% <str<strong>on</strong>g>for</str<strong>on</strong>g> three classes, showing the importance of this modality<str<strong>on</strong>g>for</str<strong>on</strong>g> at least <strong>on</strong>e participant. However there is an excepti<strong>on</strong>, the CE classificati<strong>on</strong> task, where theLDA does not have the best accuracy. In this task the sparse kernel machines have betteraccuracies but they were sensitive to the unbalanced nature of this c<strong>on</strong>figurati<strong>on</strong> with 200samples bel<strong>on</strong>ging to the excited class <strong>and</strong> 100 samples bel<strong>on</strong>ging to the calm class. As can beseen from the c<strong>on</strong>fusi<strong>on</strong> matrices of Table 6.2, sparse kernel machines tend to always assign theexcited class to test samples. Those results where thus c<strong>on</strong>sidered as irrelevant <strong>and</strong> the LDAclassifier chosen as the most relevant classifier <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong>.(a)ClassifiedClassifiedCalm ExcitedCalm ExcitedTruthTruth(b)Calm 65% 35% Calm 31% 69%Excited 33% 67% Excited 11% 89%ClassifiedClassifiedTruthCalm ExcitedTruthCalm Excited(c)(d)Calm 13% 87% Calm 33% 67%Excited 3% 97% Excited 10% 90%Table 6.2. Average c<strong>on</strong>fusi<strong>on</strong> matrices across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features <strong>and</strong> different classifiers: (a)LDA, (b) Linear SVM, (c) RBF SVM <strong>and</strong> (d) Linear RVM.Compared to the state of the art of emoti<strong>on</strong> assessment from peripheral signals <strong>and</strong> time segmentsof similar durati<strong>on</strong> our results are under those reported. In [107] 90% <strong>and</strong> 97% of accuracy wasobtained using time windows of 2 s <str<strong>on</strong>g>for</str<strong>on</strong>g> valence <strong>and</strong> arousal assessment respectively. However,the accuracy they report represents the number of samples from which the output of a neuralnetwork regressor falls in a 20% interval of the target value. Thus this accuracy cannot be directlycompared to classificati<strong>on</strong> tasks. In [114] the classificati<strong>on</strong> strategy discriminated three emoti<strong>on</strong>alstates (neutral, positive <strong>and</strong> negative) with an accuracy of 71% from 6 s signals. Theclassificati<strong>on</strong> strategy used in [114] included a detecti<strong>on</strong> of signals corrupti<strong>on</strong>, whichdem<strong>on</strong>strates the importance of such a procedure <str<strong>on</strong>g>for</str<strong>on</strong>g> correct emoti<strong>on</strong> assessment. However, theaccuracy was computed <str<strong>on</strong>g>for</str<strong>on</strong>g> <strong>on</strong>ly <strong>on</strong>e participant after training the algorithm <strong>on</strong> 8 participants. Togive an example of the variability of results that can be obtained from a participant to another, inour study results ranged from 39% <str<strong>on</strong>g>for</str<strong>on</strong>g> the worst participant to 78% <str<strong>on</strong>g>for</str<strong>on</strong>g> the best c<strong>on</strong>sidering <strong>on</strong>lythe LDA classifier (the classifier having the lowest variance) <strong>and</strong> the CPN classificati<strong>on</strong> task.112


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>sThe large differences in accuracy between the EEG <strong>and</strong> peripheral features can be explained bytwo factors. Firstly, the protocol is based <strong>on</strong> a cognitive elicitati<strong>on</strong> of emoti<strong>on</strong>s where participantsare asked to remember past emoti<strong>on</strong>al episodes which ensures str<strong>on</strong>g brain activities. Moreover,the emphasis was put <strong>on</strong> the internal feeling of emoti<strong>on</strong>s rather than <strong>on</strong> the expressi<strong>on</strong> of emoti<strong>on</strong>that can help to induce peripheral reacti<strong>on</strong>s [5]. Sec<strong>on</strong>dly, the 8 s length of trials may be too short<str<strong>on</strong>g>for</str<strong>on</strong>g> a complete activati<strong>on</strong> of peripheral signals while it may be sufficient <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG signals.For both EEG <strong>and</strong> peripheral features there is always high variability of results across theparticipants. For instance the accuracies ranged from 45% to 86% to classify emoti<strong>on</strong>s in threeclasses using the EEG_STFT features <strong>and</strong> the Linear SVM classifier. This variability can beexplained by the fact that the participants had more or less difficulty in accomplishing therequested tasks as reported during the interview. Another remark that holds <str<strong>on</strong>g>for</str<strong>on</strong>g> all feature sets isthat the detecti<strong>on</strong> of arousal states is more accurate than the detecti<strong>on</strong> of valence sates. This is notsurprising <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral activity since it is known to better correlate with the arousal axis thanwith the valence axis [7], <strong>and</strong> sheds some new light <strong>on</strong> the usability of EEG <str<strong>on</strong>g>for</str<strong>on</strong>g> the detecti<strong>on</strong> ofarousal <strong>and</strong> valence. Notice that the st<strong>and</strong>ard deviati<strong>on</strong> is lower <str<strong>on</strong>g>for</str<strong>on</strong>g> arousal identificati<strong>on</strong> than <str<strong>on</strong>g>for</str<strong>on</strong>g>all other combinati<strong>on</strong> of emoti<strong>on</strong>al states showing that arousal states are detected with morestability across participants. The RVM classifier also tends to classify emoti<strong>on</strong> with morestability but un<str<strong>on</strong>g>for</str<strong>on</strong>g>tunately obtained a lower average accuracy than SVM <str<strong>on</strong>g>for</str<strong>on</strong>g> most classificati<strong>on</strong>schemes.This study also allows comparing the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mances of the different classifiers in the three featurespaces. For the Peripheral feature set (Figure 6.5), the classifiers have relatively similaraccuracies except <str<strong>on</strong>g>for</str<strong>on</strong>g> the QDA which per<str<strong>on</strong>g>for</str<strong>on</strong>g>ms poorly compared to the others. Since thisalgorithm needs to compute a covariance matrix <str<strong>on</strong>g>for</str<strong>on</strong>g> each class, the low number of samples thatare available <str<strong>on</strong>g>for</str<strong>on</strong>g> learning (around 100 per class) explains this result. The RBF SVM does notper<str<strong>on</strong>g>for</str<strong>on</strong>g>m as well as the other classifiers <str<strong>on</strong>g>for</str<strong>on</strong>g> the two classes <str<strong>on</strong>g>for</str<strong>on</strong>g>mulati<strong>on</strong>s, suggesting that thoseproblems are linear by nature. For the high dimensi<strong>on</strong>al spaces of EEG features the LDAaccuracy is always about 10% below the results obtained by SVM classificati<strong>on</strong>. This c<strong>on</strong>firmsthe effectiveness of SVM’s in high dimensi<strong>on</strong>al spaces [138]. One of the goals of the presentwork was also to determine which of the RVM <strong>and</strong> probabilistic SVM would have the bestaccuracies in order to use the best algorithm <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of fusi<strong>on</strong>. As can be seen fromFigure 6.3 <strong>and</strong> Figure 6.4, the probabilistic SVM per<str<strong>on</strong>g>for</str<strong>on</strong>g>ms as well as the st<strong>and</strong>ard SVMdem<strong>on</strong>strating the interest of such a classifier to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m fusi<strong>on</strong> <strong>on</strong> the basis of st<strong>and</strong>ardizedscores. The RVM classifier outper<str<strong>on</strong>g>for</str<strong>on</strong>g>ms the LDA, showing its adequacy <str<strong>on</strong>g>for</str<strong>on</strong>g> high dimensi<strong>on</strong>alspaces but does not outper<str<strong>on</strong>g>for</str<strong>on</strong>g>m the SVM. An explanati<strong>on</strong> could be that RVM’s generally usedless support vectors than SVM’s which is not desirable in those undersampled classificati<strong>on</strong> taskswhere good generalizati<strong>on</strong> is hard to obtain.113


Chapter 66.4.3 Results of feature selecti<strong>on</strong>Feature selecti<strong>on</strong> was applied <str<strong>on</strong>g>for</str<strong>on</strong>g> participant 1 <strong>on</strong> the EEG_STFT feature set. Only the datarecorded from the first participant was subject to this analysis because of the computati<strong>on</strong>al timeneeded to select features from this high dimensi<strong>on</strong>al feature set. The FCBF algorithm was applied<strong>on</strong> all classificati<strong>on</strong> schemes while the filter method based <strong>on</strong> the Fisher criteri<strong>on</strong> was applied<strong>on</strong>ly the CPN classificati<strong>on</strong> scheme.a. FCBFFor the FCBF feature selecti<strong>on</strong>, Figure 6.6 shows classificati<strong>on</strong> accuracies as a functi<strong>on</strong> of thethreshold FCBF . For LDA, accuracies <str<strong>on</strong>g>for</str<strong>on</strong>g> all two-class sets without feature selecti<strong>on</strong> wereobtained using a subset of the original feature set, generally with FCBF =0.2 <strong>and</strong> a subset size of30 to 80 features. FCBF even succeeded in improving the results <str<strong>on</strong>g>for</str<strong>on</strong>g> about 6% <str<strong>on</strong>g>for</str<strong>on</strong>g> the CNclassificati<strong>on</strong> task. The significant reducti<strong>on</strong> of the number of features (less than 0.5% of originalfeatures are kept) can help improve computati<strong>on</strong> time <strong>and</strong> storage capacity in a practicalapplicati<strong>on</strong>.The number of features that have a correlati<strong>on</strong> with the class labels lower than 0.05 representsaround half of the original size of the EEG_STFT feature set. However <str<strong>on</strong>g>for</str<strong>on</strong>g> all classificati<strong>on</strong>schemes the FCBF algorithm keeps <strong>on</strong>ly around 100 of those features. This is due to the sec<strong>on</strong>dstep of the FCBF algorithm that removes redundant features <strong>and</strong> shows that many features arecorrelated. This correlati<strong>on</strong> is certainly due to the c<strong>on</strong>structi<strong>on</strong> of the EEG_STFT feature set: twoenergy features extracted from the same electrode signal but in a neighboring time window arelikely to be correlated, as are two features computed over the same time window <strong>and</strong> from thesignals of two neighboring electrodes.SVM accuracy is higher without feature selecti<strong>on</strong>, c<strong>on</strong>firming its intrinsic capacity of goodgeneralizati<strong>on</strong> in high dimensi<strong>on</strong>al spaces. For both LDA <strong>and</strong> SVM feature selecti<strong>on</strong> was noteffective <strong>on</strong> the set of three classes (42% of accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> <strong>on</strong>e selected feature <strong>and</strong> up to 47% <str<strong>on</strong>g>for</str<strong>on</strong>g> 6selected features). This can be due to the n<strong>on</strong>linear nature of this problem since the FCBFalgorithm relies <strong>on</strong> linear correlati<strong>on</strong>. One soluti<strong>on</strong> can be to substitute linear correlati<strong>on</strong> bymutual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> as a measure of relevance [149]. Another explanati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the accuracydecrease when using feature selecti<strong>on</strong> is that the FCBF algorithms eliminates features withouttaking into account the quality of the whole feature subset as is the case with wrapper algorithms.FCBF then fails to find features that are interacting <strong>and</strong> such that they are improvingclassificati<strong>on</strong> accuracy with the complete set of features.114


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>sAccuracyAccuracyCPN CE NP1106 106 90 34 6 11102 102 91 38 6 10.90.90.80.80.70.70.60.60.50.50.40.40.30.30.20.20.10.100.05 0.1 0.15 0.2 0.25 0.300.05 0.1 0.15 0.2 0.25 0.31103 103 102 85 26 80.90.80.70.60.50.40.30.20.100.05 0.1 0.15 0.2 0.25 0.3 FCBF FCBF FCBFCP107 103 84 35 91100 100 99 80 28 310710.90.90.80.80.7LDA - FCBF0.70.6LDA - no feature selecti<strong>on</strong>0.60.50.5Linear SVM - FCBF0.40.40.3Linear SVM - no feature selecti<strong>on</strong> 0.30.20.20.10.1000.05 0.1 0.15 0.2 0.25 0.30.05 FCBF0.1 0.15 0.2 FCBF0.25 0.3CNFigure 6.6. classificati<strong>on</strong> accuracy using participant 1 EEG_STFT features with LDA <strong>and</strong> with SVM <strong>on</strong> thefive sets of classes, with or without FCBF feature selecti<strong>on</strong>. The bottom horiz<strong>on</strong>tal axis indicates the value ofthe threshold FCBF , while the top horiz<strong>on</strong>tal axis corresp<strong>on</strong>ds to the number of selected features.b. Fisher feature selecti<strong>on</strong>Since the FCBF algorithm has a poor per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance <strong>on</strong> the CPN classificati<strong>on</strong> scheme, the filtermethod based <strong>on</strong> the Fisher criteri<strong>on</strong> was applied <strong>on</strong> this problem <strong>and</strong> the accuracies of LDA <strong>and</strong>linear SVM classifiers were analyzed. Figure 6.7 shows the accuracy of those classifiers as afuncti<strong>on</strong> of the number of selected features, the last point of the curve being the accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> thecomplete set of 16704 features.As can be seen from by comparing Figure 6.6 (CPN classes) <strong>and</strong> Figure 6.7 the accuraciesobtained with the Fisher feature selecti<strong>on</strong> are better than those obtained with the FCBF algorithm.With Fisher selecti<strong>on</strong>, the accuracy of the LDA is 47.3% by selecting <strong>on</strong>ly <strong>on</strong>e feature <strong>and</strong> 59%by selecting 100 features compared to the best accuracy of 47% obtained with the FCBFalgorithm (the number of feature ranging from 1 to 106). The first peak in accuracy can beobserved <str<strong>on</strong>g>for</str<strong>on</strong>g> 1000 selected features using a LDA <strong>and</strong> 1500 selected features using the SVM. Thispoint corresp<strong>on</strong>ds to the best accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> the LDA <strong>and</strong> dem<strong>on</strong>strates the usefulness of featureselecti<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> this classifier. By keeping 1000 features <strong>on</strong>ly 6% of original features are selectedwhich is interesting <str<strong>on</strong>g>for</str<strong>on</strong>g> computati<strong>on</strong>al complexity <strong>and</strong> storage improvement. However, thenumber of features is still high at this point which shows that many features from the STFTfeature set are required <str<strong>on</strong>g>for</str<strong>on</strong>g> accurate detecti<strong>on</strong> of emoti<strong>on</strong>s.115


Chapter 60.70.65Accuracy0.60.55Figure 6.7. Accuracy of LDA <strong>and</strong> the Linear SVM classifiers <str<strong>on</strong>g>for</str<strong>on</strong>g> different numbers of selected features of theEEG_STFT feature set using the Fisher criteri<strong>on</strong> (<strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g> the CPN classificati<strong>on</strong> scheme). Only the numberof features marked with a ‘+’ have been computed while the other values are linearly interpolated.Figure 6.7 shows that the SVM accuracy is lower than the LDA accuracy in low dimensi<strong>on</strong>alspaces (under approximately 1250 selected features) while it is better in high dimensi<strong>on</strong>al spaces.The SVM per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance in high dimensi<strong>on</strong>al spaces can be explained by its intrinsic properties[138]. The good per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of the LDA can be explained by the fact that the diag<strong>on</strong>alizedversi<strong>on</strong> of the LDA was employed. Since in this classifier <strong>on</strong>ly the variance <strong>and</strong> the mean of thefeatures are taken into account to determine the linear boundary between the classes (c<strong>on</strong>trarily tothe SVM with a linear kernel); the diag<strong>on</strong>alized LDA is very close to the Fisher criteri<strong>on</strong>employed <str<strong>on</strong>g>for</str<strong>on</strong>g> feature selecti<strong>on</strong>.The decrease of accuracy observed after the first peak is likely due to the additi<strong>on</strong> of noisy <strong>and</strong>redundant features. However, after the additi<strong>on</strong> of more than 8000 features the accuracy increaseswith the number of selected features. This dem<strong>on</strong>strates that less relevant features (in the sense ofthe Fisher criteri<strong>on</strong>) can interact with others to improve the accuracy.As can be seen from Secti<strong>on</strong> 6.4.3.a <strong>and</strong> Secti<strong>on</strong> 6.4.3.b, feature selecti<strong>on</strong> can improve theaccuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> the LDA with a str<strong>on</strong>g reducti<strong>on</strong> of the number of features (from 0.5% to 6% of thefeatures are kept), providing advantages <str<strong>on</strong>g>for</str<strong>on</strong>g> real applicati<strong>on</strong>s where the issues of storage <strong>and</strong>computati<strong>on</strong> time are important. However, the methods used <str<strong>on</strong>g>for</str<strong>on</strong>g> feature selecti<strong>on</strong> were not able toincrease the best accuracy obtained by the SVM classifier. For this reas<strong>on</strong>, feature selecti<strong>on</strong>algorithms were not employed <str<strong>on</strong>g>for</str<strong>on</strong>g> the next steps that are fusi<strong>on</strong> of the modalities <strong>and</strong> rejecti<strong>on</strong> ofsamples.1160.56.4.4 Results of fusi<strong>on</strong>LDALinear SVM0.450 2000 4000 6000 8000 10000 12000 14000 16000 18000Number of selected featuresC<strong>on</strong>cerning the fusi<strong>on</strong> by c<strong>on</strong>catenati<strong>on</strong> of the feature vectors, no significant improvement of theaccuracy was observed. Most of the time, the classificati<strong>on</strong> accuracy was the same as the


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>saccuracy obtained <strong>on</strong> the EEG_STFT which have the highest number of features. This is likelydue to the problem of high dimensi<strong>on</strong>ality of this space. While applying feature selecti<strong>on</strong>methods after or be<str<strong>on</strong>g>for</str<strong>on</strong>g>e fusi<strong>on</strong> can help to solve this issue, this method was not analyzed in viewof the results presented in Secti<strong>on</strong> 6.4.3. Instead the fusi<strong>on</strong> at the classifier level was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med.Fusi<strong>on</strong> of classifier decisi<strong>on</strong>s is d<strong>on</strong>e according to the explanati<strong>on</strong> given in Secti<strong>on</strong> 6.3.4.According to the obtained results, fusi<strong>on</strong> was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med choosing probabilistic SVM as theclassifiers <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG features sets (q SFFT <strong>and</strong> q MI ), <strong>and</strong> the LDA as the classifier <str<strong>on</strong>g>for</str<strong>on</strong>g> the peripheralfeature set (q Periph ).Results from the fusi<strong>on</strong> of MI <strong>and</strong> STFT EEG features as well as fusi<strong>on</strong> of all EEG <strong>and</strong>peripheral features are presented in Figure 6.8. As can be seen, combining EEG feature setsincreased the best average accuracy by 2% to 4% while combining the three feature setsincreased it by 3% to 7% depending <strong>on</strong> the classificati<strong>on</strong> scheme. For instance, the accuracy ofthe SVM classifier is 63.5% <str<strong>on</strong>g>for</str<strong>on</strong>g> the CPN classificati<strong>on</strong> scheme <strong>and</strong> reaches 70% after fusi<strong>on</strong> ofthe three feature sets. In all the present cases combining feature sets leads to an increase inaverage accuracy, even when fusing modalities with low accuracies such as the peripheralsignals. This dem<strong>on</strong>strates the importance of combining multiple sources of in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> fromboth the central <strong>and</strong> peripheral nervous system in emoti<strong>on</strong> detecti<strong>on</strong> from physiological signals.Accuracy100%90%80%70%60%50%40%30%20%10%0%CPN CE NP CN CPPeripheral (LDA)EEG_MI (Prob.linear SVM)EEG_STFT (Prob.linear SVM)EEG_MI +EEG_STFTPeripheral +EEG_MI +EEG_STFTFigure 6.8. Average accuracy across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> different modalities <strong>and</strong> their associated classifiers, aswell as <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong> of the two EEG <strong>and</strong> the three physiological modalities.To our knowledge there <strong>on</strong>ly is <strong>on</strong>e study that tried to fuse peripheral <strong>and</strong> EEG in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> at thefeature level [113]. In this study the authors found that the fusi<strong>on</strong> did not improve accuracycompared to EEG classificati<strong>on</strong>. The same fusi<strong>on</strong> was also analyzed in Chapter 5 where anincrease was reported <strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g> some classifiers <strong>and</strong> sets of classes. Since poor results <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong> atthe feature level were also obtained in the current study, we believes that <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of117


Chapter 6emoti<strong>on</strong> assessment the fusi<strong>on</strong> between the peripheral <strong>and</strong> the EEG modalities should be operatedat the classifier level. This is especially true when the fusi<strong>on</strong> involves feature spaces that are ofhigh dimensi<strong>on</strong>ality.6.4.5 Results of rejecti<strong>on</strong>Finally, samples with low c<strong>on</strong>fidence values are rejected using the method described in Secti<strong>on</strong>4.4 <strong>and</strong> the corresp<strong>on</strong>ding increase in accuracy is analyzed in Figure 6.9. In this figure, <strong>on</strong>ly theresults of the CPN c<strong>on</strong>figurati<strong>on</strong> are presented <str<strong>on</strong>g>for</str<strong>on</strong>g> the trials of all 10 participants (3000 trials) <strong>and</strong>different values of the reject threshold. As can be seen from Figure 6.9, no samples are rejecteduntil reject reaches the value of 33%, which is normal since max g cannot be inferior to 33%(there is the c<strong>on</strong>straintKi1k 1). The number of rejected samples that are badly classified isihigher than the number of correctly classified samples until reject becomes higher than 47%. Wechoose this value to stop rejecting samples since most of the badly classified samples are rejectedat this point.iiNumberof eliminated samples3000200010000Badly classified samplesCorrectly classified samples0.4 0.5 0.6 0.7 0.8 0.9 1 reject10.500.4 0.5 0.6 0.7 0.8 0.9 1 rejectAccuracy% of eliminated samplesFigure 6.9. Relati<strong>on</strong> between the threshold value, classificati<strong>on</strong> accuracy <strong>and</strong> the amount of eliminatedsamples <str<strong>on</strong>g>for</str<strong>on</strong>g> the CPN classificati<strong>on</strong> task.This value corresp<strong>on</strong>ds to a mean accuracy across participants of 80%, thus increasing it by about10%. This is to be compared with the 70% accuracy when per<str<strong>on</strong>g>for</str<strong>on</strong>g>ming fusi<strong>on</strong> without rejecti<strong>on</strong>,but at the cost of rejecting 40% of the samples. Such high rejecti<strong>on</strong> rate could seem problematic<str<strong>on</strong>g>for</str<strong>on</strong>g> a real applicati<strong>on</strong>, but is however compensated by the short recording period needed to118


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of self-induced emoti<strong>on</strong>sper<str<strong>on</strong>g>for</str<strong>on</strong>g>m classificati<strong>on</strong> <strong>and</strong> give a decisi<strong>on</strong>. For instance if two c<strong>on</strong>secutive trials are rejected, <strong>and</strong>the third <strong>on</strong>e correctly classified the whole process would still be completed within 25 s.Similar curves were plotted <str<strong>on</strong>g>for</str<strong>on</strong>g> the NP <strong>and</strong> CE classificati<strong>on</strong> schemes. The results were quitesimilar than those obtained from the CPN scheme. Using the same value of 40% <str<strong>on</strong>g>for</str<strong>on</strong>g> thepercentage of rejected samples, the increase of accuracy was respectively of 11% <strong>and</strong> 10%,resulting in an accuracy of 89% <strong>and</strong> 92% <str<strong>on</strong>g>for</str<strong>on</strong>g> the NP <strong>and</strong> CE schemes. This shows the interest ofrejecting samples to improve classificati<strong>on</strong> accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> other classificati<strong>on</strong> schemes than theCPN as well.6.5 C<strong>on</strong>clusi<strong>on</strong>sThis chapter proposes an approach to classify emoti<strong>on</strong>s in the three main areas of the valencearousalspace by using physiological signals from both the PNS <strong>and</strong> the CNS. A protocol based<strong>on</strong> the recall of past emoti<strong>on</strong>al episodes was designed to acquire short-term emoti<strong>on</strong>al data from11 participants. From the data of 10 participants we extracted three feature sets, <strong>on</strong>e <str<strong>on</strong>g>for</str<strong>on</strong>g>peripheral signals <strong>and</strong> two high dimensi<strong>on</strong>al feature space <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG signals. Using the differentfeature sets, the accuracy of several classifiers was compared <strong>on</strong> the discriminati<strong>on</strong> of thedifferent combinati<strong>on</strong>s of three emoti<strong>on</strong>al states. The reducti<strong>on</strong> of the feature space dimensi<strong>on</strong>allywas studied <str<strong>on</strong>g>for</str<strong>on</strong>g> different feature selecti<strong>on</strong> algorithms. The fusi<strong>on</strong> of the three feature sets at theclassifier level, by combining the probabilistic outputs of classifiers, was analyzed. Finally,rejecti<strong>on</strong> of trials where the c<strong>on</strong>fidence of the resulting classificati<strong>on</strong> is low was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med. In thecase the trials with low c<strong>on</strong>fidence are those that are misclassified such rejecti<strong>on</strong> should lead toan increase of accuracy.Results showed the importance of EEG signals <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment by classificati<strong>on</strong> as theyhad better accuracy than peripheral signals <strong>on</strong> the 8 s of recorded signal. Classificati<strong>on</strong> of timefrequencyfeatures derived from the EEG signals provided an average accuracy of 63% <str<strong>on</strong>g>for</str<strong>on</strong>g> threeemoti<strong>on</strong>al classes <strong>and</strong> between 73% <strong>and</strong> 80% <str<strong>on</strong>g>for</str<strong>on</strong>g> two classes. A new set of features c<strong>on</strong>tainingthe mutual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> between each pairs of electrodes was proposed to represent the interacti<strong>on</strong>between different brain areas during emoti<strong>on</strong>al processes. The accuracies obtained with this newfeature set were <strong>on</strong>ly slightly lower than those obtained with the energy features, showing theirpotential interest <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment.Despite of their low accuracy compared to EEG features, the peripheral features were shown toincrease accuracy when fused with the EEG modality at the classifier level. Fusi<strong>on</strong> of the twodifferent EEG feature sets also increased the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of the emoti<strong>on</strong> assessment. This alsodem<strong>on</strong>strates the interest of the new feature set based <strong>on</strong> mutual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>. By fusing the threephysiological feature sets the obtained accuracy is of 70% <strong>on</strong> three classes. Finally, the rejecti<strong>on</strong>119


Chapter 6of 40% of samples having a low c<strong>on</strong>fidence value increased the accuracy to up to 80% <strong>on</strong> threeclasses.The analysis of feature selecti<strong>on</strong> methods to reduce the dimensi<strong>on</strong>ality of the EEG feature spacesshowed that those methods are effective to decrease computati<strong>on</strong>al <strong>and</strong> storage costs withoutloosing too much accuracy. However, the accuracy obtained after feature selecti<strong>on</strong> was alwayslower than the best accuracy obtained without feature selecti<strong>on</strong>.Since following the stimulus <strong>on</strong>set emoti<strong>on</strong>al processes in brain <strong>and</strong> peripheral signals areexpected to be observable at different times, the explorati<strong>on</strong> of different time resoluti<strong>on</strong>s isneeded to determine the time scales favorable to emoti<strong>on</strong>al assessment from EEG <strong>and</strong> peripheralactivity. For this purpose a protocol where the exact time of the emoti<strong>on</strong> elicitati<strong>on</strong> is knownshould be designed.The high number of electrodes used in this study is also an issue since it leads to a highdimensi<strong>on</strong>al space where classificati<strong>on</strong> is difficult <strong>and</strong> it <str<strong>on</strong>g>for</str<strong>on</strong>g>bids the use of this system <str<strong>on</strong>g>for</str<strong>on</strong>g> realapplicati<strong>on</strong>s. From the analysis of feature selecti<strong>on</strong> results, it was shown that a lot of featureswere correlated, potentially because they were extracted from neighboring electrodes. Groupingthe features extracted from close electrodes in order to keep <strong>on</strong>ly relevant in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> <strong>and</strong>remove part of the noise could also be a soluti<strong>on</strong> to reduce the size of feature spaces. The studyfrom Ansari-Asl et al. [129] that tries to select relevant electrodes, based <strong>on</strong> the data from thesame protocol, is also a step toward the reducti<strong>on</strong> of feature spaces sizes.Analysis of EEG in other elicitati<strong>on</strong> c<strong>on</strong>texts should also be per<str<strong>on</strong>g>for</str<strong>on</strong>g>med to c<strong>on</strong>firm the efficiencyof EEG features <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>al assessment in less cognitive tasks, as well as when interacting withcomputer interfaces. For HCI, the described work can also be used as a guideline to decide whichclassificati<strong>on</strong> strategy to use. Finally, while the rejecti<strong>on</strong> of n<strong>on</strong>-reliable trials has been shown toimprove accuracy, the percentage of rejected samples is high <strong>and</strong> further analysis should bec<strong>on</strong>ducted to c<strong>on</strong>firm that this rejecti<strong>on</strong> can improve the in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> transfer rate.120


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesChapter 7<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer games7.1 Introducti<strong>on</strong>: the flow theory <str<strong>on</strong>g>for</str<strong>on</strong>g> gamesThe experiments presented in Chapters 5 <strong>and</strong> 6 dem<strong>on</strong>strated the usefulness of physiologicalsignals <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment. In this chapter similar experiments were per<str<strong>on</strong>g>for</str<strong>on</strong>g>med in a c<strong>on</strong>textcloser to real HCI applicati<strong>on</strong>s. This work was d<strong>on</strong>e in collaborati<strong>on</strong> with the TECFA(Educati<strong>on</strong>al Technologies Laboratory, Faculty of Psychology, University of Geneva).Due to their capability to present in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> in an interactive <strong>and</strong> playful way, computer gameshave gathered increasing interest as tools <str<strong>on</strong>g>for</str<strong>on</strong>g> educati<strong>on</strong> <strong>and</strong> training [13]. Games are alsointeresting from a human-computer interacti<strong>on</strong> (HCI) point of view, because they are an idealground <str<strong>on</strong>g>for</str<strong>on</strong>g> the design of new ways to communicate with machines. <str<strong>on</strong>g>Affective</str<strong>on</strong>g> computing [2] hasopened the path to new types of human-computer interfaces that adapt to affective cues from theuser. As <strong>on</strong>e of the main goals of games is to provide emoti<strong>on</strong>al experiences such as fun <strong>and</strong>excitement, affective computing is a promising area of research to enhance game experiences.<str<strong>on</strong>g>Affective</str<strong>on</strong>g> in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> can be used to maintain involvement of a player by adapting gamedifficulty or c<strong>on</strong>tent to induce particular emoti<strong>on</strong>al states. For this purpose, automatic assessmentof emoti<strong>on</strong>s is m<strong>and</strong>atory <str<strong>on</strong>g>for</str<strong>on</strong>g> the game to adapt in real time to the feelings <strong>and</strong> involvement of theplayer, without interrupting his / her gaming experience (like it would be the case by usingquesti<strong>on</strong>naires). The present work thus focuses <strong>on</strong> emoti<strong>on</strong> assessment from physiological signalsin the c<strong>on</strong>text of a computer game applicati<strong>on</strong>.Games can elicit a lot of different emoti<strong>on</strong>al states but knowing all of them is not necessary tomaintain involvement in the game. Many representati<strong>on</strong>s of the player’s affective state have beenused in previous studies like anxiety, frustrati<strong>on</strong>, engagement, distress <strong>and</strong> the valence-arousalspace [17, 70]. According to emoti<strong>on</strong> <strong>and</strong> flow theories [20, 161] str<strong>on</strong>g involvement in a taskoccurs when the skills of an individual meet the challenge of a task (Figure 7.1). Too muchchallenge would raise anxiety <strong>and</strong> not enough would induce boredom. Both these situati<strong>on</strong>swould restrain the player’s ability to achieve a “flow experience”, leading to less involvement,engagement <strong>and</strong> possibly interrupti<strong>on</strong> of the game [162].In a game, the change from an emoti<strong>on</strong>al state to another can occur due to two main reas<strong>on</strong>s.First, the difficulty is increased because of progressi<strong>on</strong> in the different levels but too fastcompared to the competence increase of the player (potentially giving rise to anxiety, see Figure7.1). Sec<strong>on</strong>dly, the competence of the player has increased while the game remained at the samedifficulty (potentially giving rise to boredom). In both cases, the challenge should be corrected tomaintain a state of pleasure <strong>and</strong> involvement, showing the importance of having games that adapttheir difficulty according to the competence <strong>and</strong> emoti<strong>on</strong>s of the player. <str<strong>on</strong>g>Based</str<strong>on</strong>g> <strong>on</strong> this theory, we121


Chapter 7defined three emoti<strong>on</strong>al states of interest that corresp<strong>on</strong>ds to three well separated areas of thevalence-arousal space: boredom (negative-calm), engagement (positive-excited) <strong>and</strong> anxiety(negative-excited).Challenge / difficultyof the gameAnxiety(negative excited)Engagement(postive excited)FlowchannelBoredom(negative calm)Figure 7.1. Flow chart <strong>and</strong> the suggested automatic adaptati<strong>on</strong> to emoti<strong>on</strong>al reacti<strong>on</strong>s.This work attempts to verify the validity <strong>and</strong> usefulness of the three defined emoti<strong>on</strong>al states byusing a Tetris game where the challenge is modulated by changing the level of difficulty. Selfreportsas well as physiological activity were obtained from players by using the acquisiti<strong>on</strong>protocol described in Secti<strong>on</strong> 7.2. Using those data, three analyses were c<strong>on</strong>ducted. The first aimsat validating the applicability of the flow theory <str<strong>on</strong>g>for</str<strong>on</strong>g> games (see Secti<strong>on</strong> 7.3). In the sec<strong>on</strong>danalysis, detailed in Secti<strong>on</strong> 7.4, physiological signals were used <str<strong>on</strong>g>for</str<strong>on</strong>g> the purpose of classificati<strong>on</strong>of the different states. In this case, since <strong>on</strong>e of the goals of this study is to go towardapplicati<strong>on</strong>s, particular attenti<strong>on</strong> was paid to designing classifiers that could be used <str<strong>on</strong>g>for</str<strong>on</strong>g> anygamer without having to re-train it. The third analysis c<strong>on</strong>cerned the analysis of physiologicalsignals after the occurrence of game-over events (see Secti<strong>on</strong> 7.5).7.2 Data acquisiti<strong>on</strong>7.2.1 Acquisiti<strong>on</strong> protocolCompetence of the playerChange in player’s competenceChange in game’s difficultyAutomatic adaptati<strong>on</strong> of the game difficultyA gaming protocol was designed <str<strong>on</strong>g>for</str<strong>on</strong>g> acquiring physiological signals <strong>and</strong> gathering self-reporteddata. The Tetris game (Figure 7.2) was chosen in this experiment <str<strong>on</strong>g>for</str<strong>on</strong>g> the following reas<strong>on</strong>s: it iseasy to c<strong>on</strong>trol the difficulty of the game (speed of falling blocks); it is a widely known game sothat we could expect to gather data from players with different skill levels (which occurred, seeFigure 7.3); <strong>and</strong> it is playable using <strong>on</strong>ly <strong>on</strong>e h<strong>and</strong>, which is m<strong>and</strong>atory since the other h<strong>and</strong> isused <str<strong>on</strong>g>for</str<strong>on</strong>g> placement of some data acquisiti<strong>on</strong> sensors.122


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesTop of the tetris boardBlockAdded parameters(Hidden to theparticipants).Line 2Line 1Figure 7.2. Screen shot of the Tetris (DotNETris) game.The Tetris game used in this experiment is the DotNETris software 9 (Figure 7.2). The difficultylevels implemented in DotNETris were adapted to have a wider range of difficulties than in theoriginal game. The new levels ranged from 1 to 25 in which a block goes down <strong>on</strong>e line every tsec<strong>on</strong>ds with:-(0.12 level + 0.5)t e(7.1)where the exp<strong>on</strong>ential functi<strong>on</strong> was chosen because a small change in t is not really perceivablewhen the blocks are falling slowly while it is a significant change in difficulty if the blocks arealready falling fast. The 0.12 <strong>and</strong> 0.5 values where chosen empirically to have a “smooth”increase of the difficulty. As a c<strong>on</strong>sequence, the blocks were going down a line every 0.54sec<strong>on</strong>ds at level 1 <strong>and</strong> 0.03 sec<strong>on</strong>ds at level 25. Other modificati<strong>on</strong>s to the original DotNETrisinclude:- the possibility to play the game at a given level, without change of a level when 10 linesare eliminated;9 available at http://source<str<strong>on</strong>g>for</str<strong>on</strong>g>ge.net/projects/dotnetris/ (retrieved <strong>on</strong> 29 April 2009)123


Chapter 7- the comm<strong>and</strong> to speed up the fall of the blocks was disabled so that the participants had towait <str<strong>on</strong>g>for</str<strong>on</strong>g> the blocks to go down at the chosen speed;- playing the game <str<strong>on</strong>g>for</str<strong>on</strong>g> a given durati<strong>on</strong>. Each time the blocks reach the top of the Tetrisboard (Figure 7.2) during this durati<strong>on</strong>, a game over event was reported, the board wasautomatically cleared <strong>and</strong> the participant could c<strong>on</strong>tinue to play.20 participants (mean age: 27, 13 males, all right h<strong>and</strong>ed) took part in this study. After signingthe c<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m (Appendix A), each participant played Tetris several times to determine thegame level where he/she reported engagement. This was d<strong>on</strong>e by repeating three times thethreshold method, starting from a low level <strong>and</strong> progressively increasing it until engagement wasreported by the participant or starting from a high level <strong>and</strong> decreasing it. The average of theobtained levels was then c<strong>on</strong>sidered as the participant skill level. Depending <strong>on</strong> this skill level,three experimental c<strong>on</strong>diti<strong>on</strong>s were determined: medium c<strong>on</strong>diti<strong>on</strong> (game difficulty equal to theplayer’s skill level), easy c<strong>on</strong>diti<strong>on</strong> (lower difficulty, computed by subtracting 8 levels ofdifficulty from the player’s skill level), <strong>and</strong> hard c<strong>on</strong>diti<strong>on</strong> (higher difficulty, computed by adding8 levels). As can be seen from Figure 7.3, the participants to the study reported to be engaged atdifferent levels ranging <str<strong>on</strong>g>for</str<strong>on</strong>g> most of them from 11 to 16, c<strong>on</strong>firming that they add different Tetrisskills. One of the participants was even a “Tetris expert” with a skill level of 20.76Nb. of participants54321241011 12 13 14 15 16 17 18 19 20skill levelFigure 7.3. Histogram of the skill levels of the 20 participants.Participants were then equipped with several sensors to measure their peripheral physiologicalactivity: a GSR (Galvanic Skin Resp<strong>on</strong>se) sensor to measure skin resistance, a plethysmograph torecord relative blood pressure, a respirati<strong>on</strong> belt to estimate abdomen extensi<strong>on</strong> <strong>and</strong> a temperaturesensor to measure palmar changes in temperature. Those sensors are known to measure signalsthat are related to particular emoti<strong>on</strong>al activati<strong>on</strong>s as well as useful <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> detecti<strong>on</strong> (seeSecti<strong>on</strong> 2.2.2). In additi<strong>on</strong>, an EEG system was used to record central signaling from 14participants. In this study 19 electrodes were positi<strong>on</strong>ed <strong>on</strong> the skull of participants according tothe 10-20 system (see Secti<strong>on</strong> 3.1.1). As dem<strong>on</strong>strated in other studies, EEG’s can help to assess


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesemoti<strong>on</strong>al states <strong>and</strong> is also useful to provide an index of task engagement <strong>and</strong> workload [130-133]. All signals were recorded at a 1024Hz sampling rate using the Biosemi Active 2 acquisiti<strong>on</strong>system. The acquired signals were subsequently downsampled to 256Hz using the built-inBiosemi software [126]. This allows to keep the frequency b<strong>and</strong>s of interest <str<strong>on</strong>g>for</str<strong>on</strong>g> this study <strong>and</strong> tospeed up computati<strong>on</strong>s.Once equipped with the sensors, the participants took part in 6 c<strong>on</strong>secutive sessi<strong>on</strong>s (Figure 7.4).For each sessi<strong>on</strong> the participants had to follow 3 steps: stay calm <strong>and</strong> relax <str<strong>on</strong>g>for</str<strong>on</strong>g> <strong>on</strong>e minute, playthe Tetris game <str<strong>on</strong>g>for</str<strong>on</strong>g> 5 minutes in <strong>on</strong>e of the three experimental c<strong>on</strong>diti<strong>on</strong>s (difficulty level) <strong>and</strong>finally answer a questi<strong>on</strong>naire. The first step was useful to let the physiological signals return to abaseline level, to record a baseline activity <strong>and</strong> to provide a rest period to the participants. For thesec<strong>on</strong>d step, each experimental c<strong>on</strong>diti<strong>on</strong> was applied twice <strong>and</strong> in a r<strong>and</strong>om order to account <str<strong>on</strong>g>for</str<strong>on</strong>g>side effects of time in questi<strong>on</strong>naires <strong>and</strong> physiological data. The goal of participants was toper<str<strong>on</strong>g>for</str<strong>on</strong>g>m the highest possible score. To motivate them toward this goal, a prize of 20 CHF wasoffered to three of the participants having the highest score (the participants were divided in threegroups according to their competence). The questi<strong>on</strong>naire was composed of 30 questi<strong>on</strong>s relatedto both the emoti<strong>on</strong>s they felt <strong>and</strong> their level of involvement in the game (see Appendix D). Theanswer to each questi<strong>on</strong> was given <strong>on</strong> a 7 points Likert scale. Additi<strong>on</strong>ally, participants ratedtheir emoti<strong>on</strong>s in the valence-arousal space using SAM [73] scales.Rest periodStay calm <strong>and</strong> relaxPlay the tetris game at <strong>on</strong>e of thethree difficulty levelAnswer questi<strong>on</strong>naire1 min 5 mintimeSessi<strong>on</strong> 1HardSessi<strong>on</strong> 2EasySessi<strong>on</strong> 3HardSessi<strong>on</strong> 4MediumSessi<strong>on</strong> 5MediumSessi<strong>on</strong> 6EasyR<strong>and</strong>om permutati<strong>on</strong> of difficulties <str<strong>on</strong>g>for</str<strong>on</strong>g> each participanttimeFigure 7.4. Schedule of the protocol.7.2.2 Feature extracti<strong>on</strong>From the EEG signals the EEG_FFT feature set was computed. This feature set c<strong>on</strong>tains <str<strong>on</strong>g>for</str<strong>on</strong>g> eachelectrode the EEG signal energy in the , <strong>and</strong> b<strong>and</strong>s over the complete durati<strong>on</strong> of a trial. TheEEG_W feature is also part of the EEG_FFT feature set <strong>and</strong> is related to cognitive processes suchas workload <strong>and</strong> engagement. This last feature is particularly interesting <str<strong>on</strong>g>for</str<strong>on</strong>g> this protocol sincethe participants are expected to be more engaged at the medium difficulty than at the two others.The EEG_FFT feature set c<strong>on</strong>tains a total of 319 1 58 features (3 frequency b<strong>and</strong>s <strong>and</strong> 19electrodes plus the EEG_W feature).125


Chapter 7C<strong>on</strong>cerning the peripheral activity, features extracted from the corresp<strong>on</strong>ding signals are given inTable 7.1. A detailed explanati<strong>on</strong> of the features can be found in Secti<strong>on</strong> 3.3.1 <str<strong>on</strong>g>for</str<strong>on</strong>g> the st<strong>and</strong>ardfeatures. The “Reference” column indicates the secti<strong>on</strong> of Chapter 3 that corresp<strong>on</strong>ds to anexplanati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the given advanced features. The peripheral features were computed from thecomplete durati<strong>on</strong> of the 5 minute trial <strong>and</strong> c<strong>on</strong>catenated in a feature vector. In this protocol, thedurati<strong>on</strong> of a trial was sufficiently l<strong>on</strong>g to allow <str<strong>on</strong>g>for</str<strong>on</strong>g> the computati<strong>on</strong> of peripheral features (i.e.LFfHR,HFfHR,LF / HFfHR) that require signals of l<strong>on</strong>ger durati<strong>on</strong> than the features computed in theprotocols of Chapter 5 <strong>and</strong> 6. The HR signal was computed from the BVP signal of theplethysmograph as discussed in Secti<strong>on</strong> 3.2.4 <strong>and</strong> a-posteriori correcti<strong>on</strong> of the falsely detectedbeats was applied. TheRatefRespfeature was computed from the respirati<strong>on</strong> spectrum obtained byusing the Welch algorithm (Matlab pwelch functi<strong>on</strong>) with 20 sec<strong>on</strong>ds windows <strong>and</strong> 10 sec<strong>on</strong>dsoverlap. For a given trial, all the peripheral features were c<strong>on</strong>catenated in a unique feature vectorc<strong>on</strong>taining a total of 18 features.Peripheral signal St<strong>and</strong>ard features Advanced features ReferencexxxGSR X X DecRate DecTimefGSR, fGSR, Secti<strong>on</strong> 3.3.2.bNbPeaksfGSRBVP X XHeart Rate (HR) X X X LFfHR,Respirati<strong>on</strong> X RatefResp,HF /fHR,LF HFHRDRf Secti<strong>on</strong> 3.3.2.eRespf Secti<strong>on</strong> 3.3.2.dTemperature X XTable 7.1. The features extracted from the peripheral signals.In this study the collected data are not analyzed <str<strong>on</strong>g>for</str<strong>on</strong>g> each participant separately but as a whole. Itis thus important to account <str<strong>on</strong>g>for</str<strong>on</strong>g> inter-participant variability in physiological activity (see Secti<strong>on</strong>3.3). For this reas<strong>on</strong>, the physiological signals acquired during the rest period were used tocompute a baseline activity <str<strong>on</strong>g>for</str<strong>on</strong>g> each sessi<strong>on</strong> (6 baselines per participant) that was subtracted fromthe corresp<strong>on</strong>ding physiological features. Depending <strong>on</strong> the feature, the following baselinestrategies were applied (see Table 7.2):- no baseline was subtracted <str<strong>on</strong>g>for</str<strong>on</strong>g> the features that are already a relative measure ofphysiological activity such as x<strong>and</strong>DecRatefGSR;- the last value of the baseline signal was subtracted (L) from some of the xfeatures; inthat case a xfeatures represents the signal average change from the end of the baseline;126


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesFeaturename- the method used to compute the subtracted baseline activity is the same as the <strong>on</strong>e usedBaselinestrategyFeaturename<str<strong>on</strong>g>for</str<strong>on</strong>g> the feature computati<strong>on</strong>. For instance the EEG energy in the b<strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the Fzelectrode was computed from the baseline signals <strong>and</strong> subtracted to the corresp<strong>on</strong>dingEEG feature.GSRGSRDecRatef GSRDecTimef GSRNbPeaksfGSRBVPBVPHRHRL - - - - L F F F -LFf HRHFf HRLF / HFfHRRespRatef RespDRfRespTempTempEEGfeaturesBaselineF F F F F - L - FstrategyTable 7.2. The baseline subtracti<strong>on</strong> strategy used <str<strong>on</strong>g>for</str<strong>on</strong>g> each feature. -: no subtracti<strong>on</strong> of a baseline. L: last valueof the baseline signal subtracted. F: the baseline is computed using the same method than <str<strong>on</strong>g>for</str<strong>on</strong>g> featurecomputati<strong>on</strong>.7.3 Analysis of questi<strong>on</strong>naires <strong>and</strong> of physiological featuresIn this secti<strong>on</strong> the data gathered from the questi<strong>on</strong>naires <strong>and</strong> from the computed physiologicalfeatures is analyzed to c<strong>on</strong>trol the applicability of the flow theory <str<strong>on</strong>g>for</str<strong>on</strong>g> games. For this purpose thevalidity of the following two hypotheses were tested:- H1: playing in the three different c<strong>on</strong>diti<strong>on</strong>s (difficulty levels) will give rise to differentemoti<strong>on</strong>al states;- H2: as the skill increases, the player will switch from an engagement state to a boredomstate (see Figure 7.1).7.3.1 Elicited emoti<strong>on</strong>sa. Questi<strong>on</strong>nairesTo test <str<strong>on</strong>g>for</str<strong>on</strong>g> hypothesis H1, a factor analysis was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med <strong>on</strong> the questi<strong>on</strong>naires to find the axesof maximum variance. The first two comp<strong>on</strong>ents obtained from the factor analysis account <str<strong>on</strong>g>for</str<strong>on</strong>g>55.6% of the questi<strong>on</strong>naire variance <strong>and</strong> were found to be associated with higher eigenvaluesthan the other comp<strong>on</strong>ents (the eigenvalues of the first 3 comp<strong>on</strong>ents are 10.2, 8.2 <strong>and</strong> 1.7). Thequesti<strong>on</strong>naire answers given <str<strong>on</strong>g>for</str<strong>on</strong>g> each sessi<strong>on</strong> were then projected in the new space <str<strong>on</strong>g>for</str<strong>on</strong>g>med by thetwo comp<strong>on</strong>ents <strong>and</strong> an ANOVA test was applied to those new variables to check <str<strong>on</strong>g>for</str<strong>on</strong>g> differencesin distributi<strong>on</strong>s of judgment <str<strong>on</strong>g>for</str<strong>on</strong>g> the different c<strong>on</strong>diti<strong>on</strong>s. By looking at the weights of the twocomp<strong>on</strong>ents (see Appendix D) it was found that:HR127


Chapter 7- the first comp<strong>on</strong>ent was positively correlated with questi<strong>on</strong>s related to pleasure,amusement, interest <strong>and</strong> motivati<strong>on</strong>;- the sec<strong>on</strong>d comp<strong>on</strong>ent was positively correlated with questi<strong>on</strong> corresp<strong>on</strong>ding to levels ofexcitati<strong>on</strong> <strong>and</strong> pressure <strong>and</strong> negatively correlated with calm <strong>and</strong> c<strong>on</strong>trol levels.Mean <strong>and</strong> st<strong>and</strong>ard deviati<strong>on</strong>1 st comp.Easy diff.1 st comp.Med. diff.1 st comp.Hard diff.2 nd comp.Easy diff.2 nd comp.Med. diff.2 nd comp.Hard diff.Figure 7.5. Mean <strong>and</strong> st<strong>and</strong>ard deviati<strong>on</strong> of judgments <str<strong>on</strong>g>for</str<strong>on</strong>g> each axis of the two comp<strong>on</strong>ent (comp.) space <strong>and</strong>the different difficulties (diff.): easy, medium (med.) <strong>and</strong> hard.The ANOVA test, applied <strong>on</strong> the data projected <strong>on</strong> the first comp<strong>on</strong>ent (see Figure 7.5), showedthat participants felt lower pleasure, amusement, interest <strong>and</strong> motivati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the easy <strong>and</strong> hardc<strong>on</strong>diti<strong>on</strong>s than <str<strong>on</strong>g>for</str<strong>on</strong>g> the medium <strong>on</strong>e (F=46, p


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer games<strong>and</strong> hard c<strong>on</strong>diti<strong>on</strong>. If a difference is significant (p-value < 0.1) the trend of the mean from ac<strong>on</strong>diti<strong>on</strong> to another is reported in Table 7.3.Feature F-value p-value Trend ofthe meanGSR4.4 0.01 HR 2.7 0.07 GSRFeature F-value p-value Trend ofthe mean 3.4 0.04 LFf 2.4 0.09HRDecRatefGSR3.1 0.05 Resp 5.8 < 0.01 DecTimefGSR6.7 < 0.01 Temp9.4 < 0.01 f 18.3 < 0.01 Temp10 < 0.01 NbPeaksGSRTable 7.3. F-values <strong>and</strong> p-values of the ANOVA tests applied <strong>on</strong> the peripheral features <str<strong>on</strong>g>for</str<strong>on</strong>g> the 3 difficultylevels. Only the relevant features are presented (p-value < 0.1). The “Trend of the mean” column indicates thedifferences between two c<strong>on</strong>diti<strong>on</strong>s. For instance indicate a significant decrease of the variable from theeasy to the medium c<strong>on</strong>diti<strong>on</strong> (first ) <strong>and</strong> from the medium to the hard c<strong>on</strong>diti<strong>on</strong> (sec<strong>on</strong>d ), while indicate no significant differences between the easy <strong>and</strong> medium c<strong>on</strong>diti<strong>on</strong> <strong>and</strong> a significant increase to thehard c<strong>on</strong>diti<strong>on</strong>.The decrease observed <str<strong>on</strong>g>for</str<strong>on</strong>g> the GSR, GSR,DecRatefGSRfeatures <strong>and</strong> the increase of theNbPeaksfGSRbetween the easy <strong>and</strong> medium c<strong>on</strong>diti<strong>on</strong>s indicate an increase of EDA when progressing from theeasy to the medium difficulty level. Between the easy <strong>and</strong> medium c<strong>on</strong>diti<strong>on</strong>s a significantdecrease of temperature is also observed. Those results are in favor of an increase of arousalbetween the easy <strong>and</strong> the medium c<strong>on</strong>diti<strong>on</strong>. When analyzing the GSR features changes betweenthe medium c<strong>on</strong>diti<strong>on</strong> <strong>and</strong> the hard c<strong>on</strong>diti<strong>on</strong>, <strong>on</strong>ly theDecTimefGSRfeature (percentage of negativesamples in the GSR derivative) is significantly increasing. An increase of mean HR <strong>and</strong> adecrease of temperature are also observed between the same c<strong>on</strong>diti<strong>on</strong>s. Those results suggestthat there is also an increase of arousal between the medium <strong>and</strong> hard c<strong>on</strong>diti<strong>on</strong>s but to a lesserextent than between the easy <strong>and</strong> medium c<strong>on</strong>diti<strong>on</strong>s. In summary, an increased arousal isobserved <str<strong>on</strong>g>for</str<strong>on</strong>g> increasing game difficulty, supporting the results obtained from the analysis of thequesti<strong>on</strong>naires.As can be seen from Table 7.3 a total of ten features were found to have significantly differentdistributi<strong>on</strong>s am<strong>on</strong>g the three difficulties. This suggests that the c<strong>on</strong>diti<strong>on</strong>s corresp<strong>on</strong>d todifferent emoti<strong>on</strong>al states <strong>and</strong> dem<strong>on</strong>strates the interest of those features <str<strong>on</strong>g>for</str<strong>on</strong>g> later classificati<strong>on</strong> ofthe three c<strong>on</strong>diti<strong>on</strong>s. One feature of particular interest isLFfHR, the HR energy in low frequencyb<strong>and</strong>s, because it has a lower value <str<strong>on</strong>g>for</str<strong>on</strong>g> the medium c<strong>on</strong>diti<strong>on</strong> than <str<strong>on</strong>g>for</str<strong>on</strong>g> the two others, showingthat this c<strong>on</strong>diti<strong>on</strong> can elicit particular peripheral activati<strong>on</strong>. This is also <strong>on</strong>e of the <strong>on</strong>ly featuresthat can help to distinguish the medium c<strong>on</strong>diti<strong>on</strong> from the two others.129


Chapter 7c. EEG featuresAn ANOVA test was also per<str<strong>on</strong>g>for</str<strong>on</strong>g>med <strong>on</strong> each EEG feature to test <str<strong>on</strong>g>for</str<strong>on</strong>g> differences between thethree c<strong>on</strong>diti<strong>on</strong>s. Table 7.4 gives a list of the EEG features that are relevant (p-value < 0.1).However, to illustrate the EEG activity we focused <strong>on</strong> the EEG_W feature since it is acombinati<strong>on</strong> of the other features <strong>and</strong> is known to be related to cognitive processes such asengagement <strong>and</strong> workload [130, 131, 133].Left electrodes Midline electrodes Right electrodesTheta b<strong>and</strong> C3, T7, P3, P7, O1 Fz, Cz F4, C4, T8, O2Beta b<strong>and</strong> Fp1, P7, O1 Cz C4, T8, P8, O2Table 7.4. List of the relevant EEG features (p-value < 0.1) given by frequency b<strong>and</strong> <strong>and</strong> electrode.Significant differences were observed <str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG_W feature between the three c<strong>on</strong>diti<strong>on</strong>s(F=5.5, p


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesquesti<strong>on</strong>naire data were analyzed using a pairwise t-test to verify that there was a decrease ofengagement from the first to the sec<strong>on</strong>d sessi<strong>on</strong>.The pairwise t-test <strong>on</strong> the variables of the questi<strong>on</strong>naire showed a significant decrease from thefirst to the sec<strong>on</strong>d medium c<strong>on</strong>diti<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> the questi<strong>on</strong>s “I had pleasure to play” (t=-1.8, p=0.09)<strong>and</strong> “I had to adapt to the interface” (t=-3, p=0.06). From peripheral signals, a decrease in thenumber of GSR peakstemperaturewas found.NbPeaksfGSR(t=-2.4, p=0.02) as well as an increase in the average ofTemp(t=2.6, p=0.02) <strong>and</strong> average of temperature derivative Temp(t=2.3, p=0.03)Those results are indicative of a decrease of arousal <strong>and</strong> pleasure while playing twice in the samec<strong>on</strong>diti<strong>on</strong>, thus supporting hypothesis H2. The result obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> the questi<strong>on</strong> “I had to adapt tothe interface” gives a cue that this decrease could be due to an increase of player’s competence.However the competence changes were not measured with other indicators to c<strong>on</strong>firm thispossibility. In any case, those results dem<strong>on</strong>strate the importance of having automatic adaptati<strong>on</strong>of game’s difficulty when the challenge of the game remains the same.7.4 Classificati<strong>on</strong> of the gaming c<strong>on</strong>diti<strong>on</strong>s using physiological signals7.4.1 Classificati<strong>on</strong> methodsAs shown in Secti<strong>on</strong> 7.3 several features computed from both the peripheral <strong>and</strong> central signalswere found to significantly differ between the three gaming c<strong>on</strong>diti<strong>on</strong>s. Moreover the participantsreported to be in a different emoti<strong>on</strong>al state <str<strong>on</strong>g>for</str<strong>on</strong>g> each of these c<strong>on</strong>diti<strong>on</strong>s. In this secti<strong>on</strong>, the nextstep is to investigate in more details the classificati<strong>on</strong> accuracy that can be expected fromemoti<strong>on</strong> assessment in gaming c<strong>on</strong>diti<strong>on</strong>s. For this purpose classificati<strong>on</strong> methods were applied<strong>on</strong> the data gathered from the gaming protocol. The ground-truth labels were defined as the threegaming c<strong>on</strong>diti<strong>on</strong>s, each <strong>on</strong>e being associated to <strong>on</strong>e of three states: boredom (easy c<strong>on</strong>diti<strong>on</strong>),engagement (medium c<strong>on</strong>diti<strong>on</strong>) <strong>and</strong> anxiety (hard c<strong>on</strong>diti<strong>on</strong>).Four classifiers were applied <strong>on</strong> this data set: a DLDA, a DQDA, a linear SVM <strong>and</strong> a SVM withRBF kernel. The diag<strong>on</strong>alized versi<strong>on</strong>s of the LDA <strong>and</strong> the QDA were employed because of thelow number of samples, which sometimes gives rise to the problem of singular covariancematrices. The participant cross-validati<strong>on</strong> method proposed in Secti<strong>on</strong> 4.1.2 was used to computethe accuracy of the classifiers. For each participant a classifier was trained using features of theother participants; accuracy was then computed by applying the trained model <strong>on</strong> thephysiological data of the tested participant. The gamma parameter of the RBF SVM was chosento maximize accuracy <strong>on</strong> the training set (see Secti<strong>on</strong> 4.1.3.c). Since the classifier is tested <strong>on</strong> thedata of participants that are not present in the training set, this method allows evaluating the131


Chapter 7per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of the classifier in the worst case where the model is not user-specific, i.e. noin<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> about the specificity of the user’s physiology is required <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment,except <str<strong>on</strong>g>for</str<strong>on</strong>g> a baseline recording of 1 min.Three feature selecti<strong>on</strong> algorithms were applied <strong>on</strong> this problem to find the features that arerelevant <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> <strong>and</strong> provide good generalizati<strong>on</strong> across participants. All thosealgorithms were applied <strong>on</strong> the training set to select features of interest <strong>and</strong> <strong>on</strong>ly the selectedfeatures were used <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> of the test set. The ANOVA feature selecti<strong>on</strong> was applied tokeep <strong>on</strong>ly those that are relevant to the class c<strong>on</strong>cept. Only the features having a p-value below0.1 were kept. The FCBF algorithm was applied to select relevant features <strong>and</strong> remove redundant<strong>on</strong>es. The FCBF parameter was set to 0.2 because (i) it was shown in Chapter 6 that this value isrelevant <str<strong>on</strong>g>for</str<strong>on</strong>g> FCBF EEG features selecti<strong>on</strong> <strong>and</strong> (ii) the number of features that have a correlati<strong>on</strong>with the classes higher than 0.2 (7 <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features <strong>and</strong> 23 <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG features) is similar tothe number of relevant features found using the ANOVA test (10 <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features <strong>and</strong> 20<str<strong>on</strong>g>for</str<strong>on</strong>g> EEG features, see Secti<strong>on</strong> 7.3.1). Finally, the SFFS algorithm (see Secti<strong>on</strong> 4.2.2) was alsoused to select features of interest, including potentially interacting features. To search <str<strong>on</strong>g>for</str<strong>on</strong>g> featuresthat have good generalizati<strong>on</strong> across participants, the accuracy of a feature subset was estimatedby computing the participant cross-validati<strong>on</strong> accuracy <strong>on</strong> the training set. The maximum sizeF SFFS of a feature subset was set to 18 <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features <strong>and</strong> 20 <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG features. We limitedthe maximum size of the EEG feature set to 20 because the SFFS algorithm is computati<strong>on</strong>allyexpensive <strong>and</strong> 20 features were found to be relevant according to the ANOVA test.Since the EEG signals were recorded <strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g> 14 out of the 20 participants, the available numberof samples <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG based classificati<strong>on</strong> is not the same as <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral based classificati<strong>on</strong>. Forthis reas<strong>on</strong> the results obtained from EEG <strong>and</strong> peripheral features are separated in two secti<strong>on</strong>swith classificati<strong>on</strong> algorithm applied <strong>on</strong> 14 participants <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG <strong>and</strong> 20 participants <str<strong>on</strong>g>for</str<strong>on</strong>g>peripheral features. In Secti<strong>on</strong> 7.4.4 the classificati<strong>on</strong> accuracies obtained with EEG <strong>and</strong>peripheral features <strong>on</strong> different time scales are compared while the fusi<strong>on</strong> of peripheral <strong>and</strong> EEGmodalities is investigated in Secti<strong>on</strong> 7.4.5. In both cases, the classificati<strong>on</strong> accuracy wascomputed <strong>on</strong>ly <strong>on</strong> the 14 participants having EEG recorded.7.4.2 Peripheral signalsFigure 7.7 presents the accuracies obtained by applying the classificati<strong>on</strong> methods <strong>on</strong> the featuresextracted from the peripheral signals. The result obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> the linear SVM is omitted <str<strong>on</strong>g>for</str<strong>on</strong>g> theSFFS. When using the SFFS algorithm to search <str<strong>on</strong>g>for</str<strong>on</strong>g> the first best feature, the computati<strong>on</strong>s couldnot be completed, this presumably was caused by c<strong>on</strong>vergence problems or by an error in thelibSVM toolbox implementati<strong>on</strong>.132


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesAccuracy90%80%70%60%50%40%30%20%10%No feat. select.ANOVAFCBFSFFS0%DLDA DQDA Linear SVM RBF SVMFigure 7.7. Accuracies of the different classifiers <strong>and</strong> feature selecti<strong>on</strong> metods <strong>on</strong> the peripheral features.Without feature selecti<strong>on</strong> the linear classifiers obtained the best accuracies of 55% <str<strong>on</strong>g>for</str<strong>on</strong>g> the linearSVM <strong>and</strong> 54% <str<strong>on</strong>g>for</str<strong>on</strong>g> the DLDA showing their ability to find a boundary that generalizes wellacross participants. In any case, the accuracies are higher than the r<strong>and</strong>om level of 33%. Except<str<strong>on</strong>g>for</str<strong>on</strong>g> the ANOVA, the feature selecti<strong>on</strong> methods always improved the classificati<strong>on</strong> accuracies.The best accuracy of 59% is obtained with the DQDA combined with SFFS feature selecti<strong>on</strong>.However the FCBF results (58%) are not significantly different from those obtained with theSFFS algorithm because of the high variance of the accuracies. Moreover, the variance of theaccuracies obtained with SFFS tends to be higher than those obtained with the FCBF whichshows that the FCBF is more stable than the SFFS algorithm in selecting the proper features.According to the results <strong>and</strong> c<strong>on</strong>sidering that the FCBF is much faster than the SFFS, the FCBFcan be c<strong>on</strong>sidered as the best feature selecti<strong>on</strong> algorithm <str<strong>on</strong>g>for</str<strong>on</strong>g> this classificati<strong>on</strong> scheme.Since the participant cross-validati<strong>on</strong> method was used, the feature selecti<strong>on</strong> algorithms wereapplied 20 times <strong>on</strong> different training sets. For this reas<strong>on</strong>, the features selected at each iterati<strong>on</strong>of the cross-validati<strong>on</strong> procedure can be different. The histograms of Figure 7.8 show <str<strong>on</strong>g>for</str<strong>on</strong>g> eachfeature the number of times it was selected by a given feature selecti<strong>on</strong> algorithm. The averagenumber of selected features is 3.5 <str<strong>on</strong>g>for</str<strong>on</strong>g> the FCBF, 9.35 <str<strong>on</strong>g>for</str<strong>on</strong>g> the ANOVA feature selecti<strong>on</strong> <strong>and</strong> 4.8<str<strong>on</strong>g>for</str<strong>on</strong>g> the SFFS. The ANOVA nearly always selected the features that were found to be relevant inSecti<strong>on</strong> 7.3.1.b but with poor resulting accuracy (Figure 7.7). Thanks to the removal of redundantfeatures, the FCBF str<strong>on</strong>gly reduces the original size of the feature space with a good resultingaccuracy. Moreover this algorithm nearly always selected the same features independently of thetraining set showing its stability. The SFFS also obtained good per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance but as can be seenfrom Figure 7.8, some of the features were selected <strong>on</strong>ly <strong>on</strong> some of the training sets, showingthat this algorithm is less stable than the FCBF.133


Chapter 7FCBFANOVASFFS Tempµ TempLF / HFf HRLFf HRHFf HR HR HRµ HR BVPµ BVPDRf Resp RespRatef RespNbPeaksf GSRDecTimef GSRDecRatef GSR GSRµ GSR0 5 10 15 200 5 10 15 200 5 10 15 20Figure 7.8. Histograms of the number of cross-validati<strong>on</strong> iterati<strong>on</strong>s (over a total of 20) in which features havebeen selected by the FCBF, ANOVA <strong>and</strong> SFFS feature selecti<strong>on</strong> algorithms. The SFFS feature selecti<strong>on</strong> isdisplayed <str<strong>on</strong>g>for</str<strong>on</strong>g> the DQDA classificati<strong>on</strong>.By inspecting the SFFS, FCBF <strong>and</strong> ANOVA selected features, theDecTimefGSR<strong>and</strong>NbPeaksfGSRfeatureswere always selected which shows their importance <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> of the three c<strong>on</strong>diti<strong>on</strong>s fromphysiological signals. To our knowledge similar features have been used <strong>on</strong>ly in [109] <str<strong>on</strong>g>for</str<strong>on</strong>g>emoti<strong>on</strong>s assessment despite of their apparent relevance. The HRfeature was frequently selectedby the FCBF but never by the SFFS <strong>and</strong> vice-versa <str<strong>on</strong>g>for</str<strong>on</strong>g> the Respfeature. The Respfeature wasremoved by the FCBF because it was correlated withHR. However the SFFS kept the Respfeature based <strong>on</strong> its predictive accuracy which suggests that this feature may be better than<str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong>. Finally, the temperature features were also found to be frequently relevant.Because of its good accuracy <strong>and</strong> low computati<strong>on</strong>al time the FCBF algorithm coupled withDQDA classificati<strong>on</strong> was used <str<strong>on</strong>g>for</str<strong>on</strong>g> further analyses involving the peripheral modality. Table 7.5presents the c<strong>on</strong>fusi<strong>on</strong> matrix <str<strong>on</strong>g>for</str<strong>on</strong>g> the 3 classes: it can be seen that the boredom c<strong>on</strong>diti<strong>on</strong> waswell classified, followed by the anxiety c<strong>on</strong>diti<strong>on</strong>. Samples from the engagement c<strong>on</strong>diti<strong>on</strong> tendto be classified mostly as bored samples <strong>and</strong> also as anxious samples. This is not surprising sincethis c<strong>on</strong>diti<strong>on</strong> lies in between the others. Notice that 21% of the samples bel<strong>on</strong>ging to the anxietyclass are classified as bored samples; this can be due to fact that some participants completelydisengaged from the task because of its difficulty, reaching an emoti<strong>on</strong>al state close to boredom.In this case, the adaptive game we propose would increase the level of difficulty since theHR134


Chapter 7betaalphathetaEEG-WO2P8P4T8C4CzF8F4FzFp2PzO1P7P3T7C3F7F3Fp1O2P8P4T8C4CzF8F4FzFp2PzO1P7P3T7C3F7F3Fp1O2P8P4T8C4CzF8F4FzFp2PzO1P7P3T7C3F7F3Fp1Figure 7.10. Histograms of the number of cross-validati<strong>on</strong> iterati<strong>on</strong>s (over a total of 14) in which features havebeen selected by the FCBF, ANOVA <strong>and</strong> SFFS feature selecti<strong>on</strong> algorithms. The SFFS feature selecti<strong>on</strong> isdisplayed <str<strong>on</strong>g>for</str<strong>on</strong>g> the DLDA classificati<strong>on</strong>.As can be seen from Figure 7.10, the FCBF selected less features than the two other featureselecti<strong>on</strong> methods. It selected 3.1 features in average compared to 20.3 <str<strong>on</strong>g>for</str<strong>on</strong>g> the ANOVA <strong>and</strong> 13.0<str<strong>on</strong>g>for</str<strong>on</strong>g> the SFFS coupled with the DLDA. This explains the low accuracy obtained with the FCBF<strong>and</strong> shows that good accuracies <strong>on</strong> this problem can be obtained <strong>on</strong>ly by c<strong>on</strong>catenating severalfeatures. The ANOVA algorithm often selected the features described in Secti<strong>on</strong> 7.3.1.c. TheSFFS coupled with the DLDA had accuracies close to those of the ANOVA with DLDA but byselecting less features in average. For this reas<strong>on</strong> the features selected by this method are ofparticular importance <str<strong>on</strong>g>for</str<strong>on</strong>g> accurate classificati<strong>on</strong> of the three gaming c<strong>on</strong>diti<strong>on</strong>s. The more oftenselected features (selected more than 8 times) were the theta b<strong>and</strong> energies of the T7, O1, Cz, P4<strong>and</strong> P3 electrodes <strong>and</strong> the beta b<strong>and</strong> energies of the P7, Pz <strong>and</strong> O2 electrodes. This result showsthat the occipital <strong>and</strong> parietal lobes were particularly useful <str<strong>on</strong>g>for</str<strong>on</strong>g> differentiati<strong>on</strong> of the three gamingc<strong>on</strong>diti<strong>on</strong>s.FCBF0 5 10 15ANOVA0 5 10 150 5 10 15The c<strong>on</strong>fusi<strong>on</strong> matrix displayed in Table 7.6 <str<strong>on</strong>g>for</str<strong>on</strong>g> the DLDA <strong>and</strong> FCBF methods shows that thedifferent classes were detected with similar accuracies. The medium c<strong>on</strong>diti<strong>on</strong> still has the lowestaccuracy but is better detected than when using the peripheral features. On the other h<strong>and</strong>, theSFFS136


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gameseasy c<strong>on</strong>diti<strong>on</strong> is detected with less accuracy than with peripheral features. This indicates that thefusi<strong>on</strong> of the two modalities should increase the overall accuracy.Estimated EasyMediumHardTrue(Boredom) (Engagement) (Anxiety)Easy (Boredom) 57% 43% 0%Medium (Engag.) 21% 50% 29%Hard (Anxiety) 19% 19% 62%Table 7.6. C<strong>on</strong>fusi<strong>on</strong> matrix <str<strong>on</strong>g>for</str<strong>on</strong>g> the DLDA classifier with ANOVA feature selecti<strong>on</strong>.7.4.4 EEG <strong>and</strong> peripheral signalsIn order to compare accuracies obtained using either EEG or peripheral signals, the bestcombinati<strong>on</strong>s of classifiers <strong>and</strong> feature selecti<strong>on</strong> methods were applied <strong>on</strong> the physiologicaldatabase with the same number of participants <str<strong>on</strong>g>for</str<strong>on</strong>g> both modalities (the 14 participants <str<strong>on</strong>g>for</str<strong>on</strong>g> whomEEG was recorded). Moreover, the comparis<strong>on</strong> was c<strong>on</strong>ducted <str<strong>on</strong>g>for</str<strong>on</strong>g> different time scales to analyzethe per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of each modality as a functi<strong>on</strong> of the signal durati<strong>on</strong> used <str<strong>on</strong>g>for</str<strong>on</strong>g> the featurescomputati<strong>on</strong>. For this purpose, each sessi<strong>on</strong> (see Figure 7.4) was divided into 1 to 10 n<strong>on</strong>overlappingwindows of 300/W sec<strong>on</strong>ds, where W is the number of windows <strong>and</strong> 300 sec<strong>on</strong>ds thedurati<strong>on</strong> of a sessi<strong>on</strong>. An EEG <strong>and</strong> peripheral feature vector was then computed from eachwindow <strong>and</strong> the label of the sessi<strong>on</strong> was attributed to this feature vector. By using this method, adatabase of physiological features was c<strong>on</strong>structed <str<strong>on</strong>g>for</str<strong>on</strong>g> each window size ranging from 30 to 300sec<strong>on</strong>ds.For a database in which the features were computed from W windows, the number of samples <str<strong>on</strong>g>for</str<strong>on</strong>g>each class is 202 W (20 participants, 2 sessi<strong>on</strong>s per class <strong>and</strong> W windows per sessi<strong>on</strong>). Thusthe number of samples per class increases with W. Since the number of samples can influenceclassificati<strong>on</strong> accuracy <strong>and</strong> the goal of this study is to analyze the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of EEG <strong>and</strong>peripheral features at different time scales, it is important that this comparis<strong>on</strong> be c<strong>on</strong>ducted withthe same number of samples <str<strong>on</strong>g>for</str<strong>on</strong>g> each window’s length. To satisfy this c<strong>on</strong>straint <strong>on</strong>e sample waschosen r<strong>and</strong>omly from each sessi<strong>on</strong> using a uni<str<strong>on</strong>g>for</str<strong>on</strong>g>m distributi<strong>on</strong> to have 202 40 samples perclass. The classificati<strong>on</strong> algorithms were than applied <strong>on</strong> this reduced database. This was repeated1000 times <str<strong>on</strong>g>for</str<strong>on</strong>g> each value of W to account <str<strong>on</strong>g>for</str<strong>on</strong>g> the different possible combinati<strong>on</strong>s of the windows(except <str<strong>on</strong>g>for</str<strong>on</strong>g> W=1). Notice that it is not possible to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m classificati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> all windowscombinati<strong>on</strong>s since there are W 40 such combinati<strong>on</strong>s.By using this method the average accuracies over the 1000 iterati<strong>on</strong>s are displayed in Figure 7.11.The small accuracy oscillati<strong>on</strong>s that can be observed <str<strong>on</strong>g>for</str<strong>on</strong>g> small time windows (less than 100sec<strong>on</strong>ds) are likely due to the increase of the number of possible combinati<strong>on</strong>s of windows. Ascan be seen from Figure 7.11 the accuracy obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> the peripheral signals with the originaldurati<strong>on</strong> of the sessi<strong>on</strong>s (300 sec<strong>on</strong>ds) is not significantly different from the <strong>on</strong>e obtained with all137


Chapter 7of the 20 of participants (See Secti<strong>on</strong> 7.4.2). Thus having 13 or 19 participants <str<strong>on</strong>g>for</str<strong>on</strong>g> classifierstraining (because of participant cross-validati<strong>on</strong>) does not significantly change the classificati<strong>on</strong>per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance. This suggests that adding more participants to the current database would notincrease classificati<strong>on</strong> accuracies <strong>and</strong> that recording 14 to 20 participants is enough to obtainreliable accuracy estimati<strong>on</strong>s.0.60.58EEG with DLDA <strong>and</strong> ANOVAPeripheral with DQDA <strong>and</strong> FCBF0.56Accuracy0.540.520.50.480.46Figure 7.11. Classificati<strong>on</strong> accuracy as a functi<strong>on</strong> of the durati<strong>on</strong> of a trial <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG <strong>and</strong> peripheral features.For both modalities, decreasing the durati<strong>on</strong> of the window <strong>on</strong> which the features are computedleads to a decrease of accuracy. However, this decrease is str<strong>on</strong>ger <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features than <str<strong>on</strong>g>for</str<strong>on</strong>g>EEG features. For the EEG features, the accuracy drops from 56% <str<strong>on</strong>g>for</str<strong>on</strong>g> windows of 300 sec<strong>on</strong>ds toaround 51% <str<strong>on</strong>g>for</str<strong>on</strong>g> windows of 30-50 sec<strong>on</strong>ds. For the peripheral features the accuracy is 57% <str<strong>on</strong>g>for</str<strong>on</strong>g>windows of 300 sec<strong>on</strong>ds <strong>and</strong> around 45% <str<strong>on</strong>g>for</str<strong>on</strong>g> windows of 30-50 sec<strong>on</strong>ds. Moreover, the EEGaccuracy remains approximately the same <str<strong>on</strong>g>for</str<strong>on</strong>g> windows having durati<strong>on</strong> inferior to 100 sec<strong>on</strong>dswhile the peripheral accuracy c<strong>on</strong>tinues to decrease. All those results dem<strong>on</strong>strate that the EEGfeatures are more robust <strong>on</strong> short term assessment than the peripheral features. For ourapplicati<strong>on</strong>, adapting the difficulty of the Tetris game based <strong>on</strong> the physiological signals gatheredduring the 5 precedent minutes may be undesirable since there is a high probability that thedifficulty of the game has changed during this laps of time due to usual game progress. Havingmodalities, like EEG, that are able to estimate the state of the user <strong>on</strong> shorter time periods is thusof great interest.1387.4.5 Fusi<strong>on</strong>0.440 50 100 150 200 250 300Window’s durati<strong>on</strong> (in sec<strong>on</strong>ds)Fusi<strong>on</strong> at the feature level (see Secti<strong>on</strong> 4.3.1) was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med <strong>and</strong> the different classifiers <strong>and</strong>feature selecti<strong>on</strong> methods were applied <strong>on</strong> the resulting feature sets. The results did not show anysignificant improvement of the classificati<strong>on</strong> accuracy. Similar results were obtained whenper<str<strong>on</strong>g>for</str<strong>on</strong>g>ming fusi<strong>on</strong> at the classifier level using the sum rule (see Secti<strong>on</strong> 4.3.2.a) <strong>and</strong> the best


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesclassifiers <str<strong>on</strong>g>for</str<strong>on</strong>g> each feature set. This is due to the fact that the posterior probabilities outputed bythe classifiers were generally higher <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG features than <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features so that the classestimated from the EEG features was often selected by the fusi<strong>on</strong>.As can be seen from the c<strong>on</strong>fusi<strong>on</strong> matrices obtained from the classificati<strong>on</strong> based <strong>on</strong> theperipheral <strong>and</strong> EEG features (Table 7.5 <strong>and</strong> Table 7.6), the errors made in with these two featuresets are quite different. As explained in Secti<strong>on</strong> 4.3.2.b the Bayes belief integrati<strong>on</strong> is well suited<str<strong>on</strong>g>for</str<strong>on</strong>g> this type of problem, <strong>and</strong> thus was employed <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong> of the best classifiers found <str<strong>on</strong>g>for</str<strong>on</strong>g> eachfeature set (the DLDA couples with ANOVA <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG features <strong>and</strong> DQDA coupled with FCBF<str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features). Another advantage of the Bayes belief integrati<strong>on</strong> is that the probabilitiesP( | yˆ) can be estimated independently <str<strong>on</strong>g>for</str<strong>on</strong>g> the two classifiers. It was thus possible to use theiqtraining data of 19 participants to compute probabilities <str<strong>on</strong>g>for</str<strong>on</strong>g> the peripheral features while <strong>on</strong>ly 13participants were used <str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG features. The resulting accuracy <strong>and</strong> c<strong>on</strong>fusi<strong>on</strong> matrices wereobtained by using the participant cross-validati<strong>on</strong> applied <strong>on</strong> the 14 participants <str<strong>on</strong>g>for</str<strong>on</strong>g> whom bothEEG <strong>and</strong> peripheral activity were recorded.Estimated EasyMediumHardTrue(Boredom) (Engagement) (Anxiety)Easy (Boredom) 82% 14% 4%Medium (Engag.) 29% 39% 32%Hard (Anxiety) 4% 27% 69%Table 7.7. C<strong>on</strong>fusi<strong>on</strong> matrix <str<strong>on</strong>g>for</str<strong>on</strong>g> the “Bayes belief integrati<strong>on</strong>” fusi<strong>on</strong>.The accuracy obtained after fusi<strong>on</strong> was 63% which corresp<strong>on</strong>ds to an increase of 5% compared tothe best accuracy obtained with the peripheral features. Table 7.7 presents the c<strong>on</strong>fusi<strong>on</strong> matrixobtained after fusi<strong>on</strong>. By comparing this table to Table 7.5 <strong>and</strong> Table 7.6 it can be observed thatthe detecti<strong>on</strong> accuracy of the easy <strong>and</strong> the hard classes was increased by 2% <strong>and</strong> 7% respectivelycompared to the accuracy obtained with the best feature set (peripheral features <str<strong>on</strong>g>for</str<strong>on</strong>g> the easy class<strong>and</strong> EEG features <str<strong>on</strong>g>for</str<strong>on</strong>g> the hard class). The accuracy obtained <strong>on</strong> the medium class with fusi<strong>on</strong>(39%) is lower than the <strong>on</strong>e obtained with EEG features (50%) but higher than with peripheralfeatures (33%). When per<str<strong>on</strong>g>for</str<strong>on</strong>g>ming classificati<strong>on</strong> based either <strong>on</strong> EEG or peripheral features manyof the hard samples were classified as easy while this problem was solved after fusi<strong>on</strong>. All theseresults dem<strong>on</strong>strate the interest of peripheral <strong>and</strong> EEG fusi<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> a more accurate detecti<strong>on</strong> of thethree c<strong>on</strong>diti<strong>on</strong>s.7.5 Analysis of game-over eventsAnalysis of the physiological signals should also be c<strong>on</strong>ducted <strong>on</strong> the basis of the events in thegame <strong>and</strong> not <strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g> the complete 5 min of a sessi<strong>on</strong> since change of emoti<strong>on</strong>al states can alsooccur during any of the sessi<strong>on</strong>s. For this purpose the analysis of physiological signals after a139


Chapter 7game-over event is per<str<strong>on</strong>g>for</str<strong>on</strong>g>med in this secti<strong>on</strong>. Notice that since the game restarted automaticallyafter each game-over event, the changes in physiological activity observed after these events canbe due either to the loss of a game or to the start of the new <strong>on</strong>e.7.5.1 MethodIn this secti<strong>on</strong>, the variati<strong>on</strong>s of the BVP, HR <strong>and</strong> GSR signals after game-over events wereanalyzed. Those signals were chosen to represent the peripheral activity because they have fastreacti<strong>on</strong> time <strong>and</strong> they are known to be indicators of arousal <strong>and</strong> mental ef<str<strong>on</strong>g>for</str<strong>on</strong>g>t [7, 163]. Since theHR signal is not sampled at regular interval <strong>and</strong> with the same sampling rate as other signals, itwas first interpolated using a cubic interpolati<strong>on</strong>. All the signals were then segmented intowindows of 5 sec<strong>on</strong>ds, each window starting at <strong>on</strong>e of the game-over triggers recorded during theacquisiti<strong>on</strong>. Each window was also associated to the label of the corresp<strong>on</strong>ding sessi<strong>on</strong> in order todistinguish between peripheral reacti<strong>on</strong>s of the different difficulty c<strong>on</strong>diti<strong>on</strong>s.During the easy sessi<strong>on</strong>s, n<strong>on</strong>e of the participants reached the top of the Tetris board (Figure 7.2)which shows that the difficulty level was appropriately chosen as easy but prevents any analysisof game-over events <str<strong>on</strong>g>for</str<strong>on</strong>g> this c<strong>on</strong>diti<strong>on</strong>. For the two other c<strong>on</strong>diti<strong>on</strong>s, the game-over events werepresent <str<strong>on</strong>g>for</str<strong>on</strong>g> all participants <strong>and</strong> occurred <strong>on</strong> average each 98 sec<strong>on</strong>ds <str<strong>on</strong>g>for</str<strong>on</strong>g> the medium c<strong>on</strong>diti<strong>on</strong><strong>and</strong> each 14 sec<strong>on</strong>ds <str<strong>on</strong>g>for</str<strong>on</strong>g> the hard c<strong>on</strong>diti<strong>on</strong>. As a result, the number of samples obtained <str<strong>on</strong>g>for</str<strong>on</strong>g> thehard c<strong>on</strong>diti<strong>on</strong> is much higher than the <strong>on</strong>e of the medium c<strong>on</strong>diti<strong>on</strong>. Due to the high frequencyof game-over events in the hard c<strong>on</strong>diti<strong>on</strong>, it is possible that an event fell in the window of theprecedent event (two game-overs within 5 sec<strong>on</strong>ds). In that case, the window associated to thefirst event was rejected.Within a given window, each signal is normalized by subtracting the amplitude of the firstsample from all samples. Once all the signals were segmented <strong>and</strong> normalized, differencesbetween the c<strong>on</strong>diti<strong>on</strong>s were investigated. For this purpose an ANOVA test was applied every 0.5sec<strong>on</strong>d <strong>on</strong> the values of each type of signal.7.5.2 ResultsFigure 7.12 shows the HR <strong>and</strong> GSR averages over the windows of each experimental c<strong>on</strong>diti<strong>on</strong>.No significant differences between the medium <strong>and</strong> hard c<strong>on</strong>diti<strong>on</strong>s were found <str<strong>on</strong>g>for</str<strong>on</strong>g> the BVPsignal. The GSR values were significantly different from 3 sec<strong>on</strong>ds after the game-over triggerwith a higher value <str<strong>on</strong>g>for</str<strong>on</strong>g> the medium c<strong>on</strong>diti<strong>on</strong> than <str<strong>on</strong>g>for</str<strong>on</strong>g> the hard c<strong>on</strong>diti<strong>on</strong>. The most significantdifference (F=5.3, p=0.02) was obtained at the fourth sec<strong>on</strong>d. For the HR signal, a significantdifference between the c<strong>on</strong>diti<strong>on</strong>s was found between 1.5 to 3 sec<strong>on</strong>ds after the game-over triggerwith a higher heart rate <str<strong>on</strong>g>for</str<strong>on</strong>g> the medium than <str<strong>on</strong>g>for</str<strong>on</strong>g> the hard c<strong>on</strong>diti<strong>on</strong>. The most significantdifference (F=6.3, p=0.01) occurred 2.5 sec<strong>on</strong>ds after game-over events.140


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesNormalized skinresistance (Ohms)200150100500-50MediumHardGSR**** **** **Normalized HR (BPM)10.50-0.5MediumHard**HR**-1000 1 2 3 4 5Time (sec<strong>on</strong>ds)-10 1 2 3 4 5Time (sec<strong>on</strong>ds)Figure 7.12.Averages of the normalized GSR <strong>and</strong> HR signals <str<strong>on</strong>g>for</str<strong>on</strong>g> the 5 sec<strong>on</strong>ds following the game-overtriggers. Points that are marked with a ‘+’ corresp<strong>on</strong>ds to the samples that were found to be significantlydifferent (p-value < 0.1) am<strong>on</strong>g the two c<strong>on</strong>diti<strong>on</strong>s. ‘**’ indicate that p-value < 0.05.The results obtained from the GSR signal indicate a higher EDA <str<strong>on</strong>g>for</str<strong>on</strong>g> the hard c<strong>on</strong>diti<strong>on</strong> than <str<strong>on</strong>g>for</str<strong>on</strong>g>the medium c<strong>on</strong>diti<strong>on</strong> which shows that the sympathetic activity was higher in the hard c<strong>on</strong>diti<strong>on</strong>.Since from the questi<strong>on</strong>naire the hard c<strong>on</strong>diti<strong>on</strong> was found to be related to higher excitati<strong>on</strong> <strong>and</strong>pressure, a potential interpretati<strong>on</strong> of this result is that the participants were more aroused <strong>and</strong>stressed because they had to start a new game that they know to be too hard relatively to theircompetences. The HR resp<strong>on</strong>ses are significantly different <strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g> a short period of time afterthe trigger. For this reas<strong>on</strong>, those resp<strong>on</strong>ses were assumed to be related to the game-over event<strong>and</strong> not to the new game. Higher HR was reported <str<strong>on</strong>g>for</str<strong>on</strong>g> unpleasant stimuli compared to pleasantstimuli in [62]. Since the participant reported higher pleasantness <strong>and</strong> amusement in the mediumc<strong>on</strong>diti<strong>on</strong> the difference in HR could be due to the decepti<strong>on</strong> of loosing a game in the mediumc<strong>on</strong>diti<strong>on</strong>.Un<str<strong>on</strong>g>for</str<strong>on</strong>g>tunately, more variables should be gathered to c<strong>on</strong>firm the interpretati<strong>on</strong>s given in theprecedent paragraph, especially self-reports c<strong>on</strong>cerning the game-over events. Nevertheless,results dem<strong>on</strong>strate that there are different patterns of peripheral activity after game-over eventsbetween the sessi<strong>on</strong>s where the participants reported higher motivati<strong>on</strong> <strong>and</strong> pleasantness <strong>and</strong>those were they reported high pressure <strong>and</strong> less motivati<strong>on</strong>. This suggests that this activity couldbe used to distinguish engaged from stressed states in games where such events occursfrequently. Examples of such games are the first-pers<strong>on</strong> shooter games where the characterdriven by the gamer is frequently “killed”. Further studies are needed to investigate if thoseresp<strong>on</strong>ses are also present <str<strong>on</strong>g>for</str<strong>on</strong>g> other games than the modified Tetris used in this protocol.141


Chapter 77.6 C<strong>on</strong>clusi<strong>on</strong>This study investigated the possible use of emoti<strong>on</strong> assessment from physiological signals toadapt the difficulty of a game. A protocol was designed to record physiological activity <strong>and</strong>gather self-reports of 20 participants playing a Tetris game at three different levels of difficulty.The difficulty levels were determined according to the competence of the players <strong>on</strong> the task.Three types of analysis were c<strong>on</strong>ducted <strong>on</strong> the data: statistical analysis of self-reports <strong>and</strong>physiological data was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med to c<strong>on</strong>trol that different cognitive <strong>and</strong> emoti<strong>on</strong>al states wereelicited by the protocol, classificati<strong>on</strong> was c<strong>on</strong>ducted to determine whether it is possible to detectthose states from physiological signals, <strong>and</strong> an analysis of the changes in physiological activityafter game-over events was per<str<strong>on</strong>g>for</str<strong>on</strong>g>med.The results obtained from the analysis of self-reports <strong>and</strong> physiological data showed that playingthe Tetris game at different levels of difficulty gave rise to different emoti<strong>on</strong>al states. The easydifficulty was related to a state of low pleasure, low pressure, low arousal <strong>and</strong> low motivati<strong>on</strong>which was determined as boredom. The medium difficulty elicited higher arousal than the easydifficulty, as well as higher pleasure, higher motivati<strong>on</strong> <strong>and</strong> higher amusement. It was thusdefined as engagement. Finally the hard difficulty was associated to anxiety since it elicited higharousal, high pressure <strong>and</strong> low pleasure. Moreover, the analysis of c<strong>on</strong>secutive engaged trialsshowed that the engagement of a player can decrease if the game difficulty does not change.These results dem<strong>on</strong>strate the importance of adapting the game difficulty according to theemoti<strong>on</strong>s of the player in order to maintain his / her engagement.The classificati<strong>on</strong> accuracy of EEG <strong>and</strong> peripheral signals to recover the three states elicited bythe gaming c<strong>on</strong>diti<strong>on</strong>s were analyzed <str<strong>on</strong>g>for</str<strong>on</strong>g> different classifiers, feature selecti<strong>on</strong> methods <strong>and</strong>durati<strong>on</strong>s <strong>on</strong> which the features were computed. Without feature selecti<strong>on</strong> the best classifiersobtained an accuracy around 55% <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features <strong>and</strong> 48% <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG features. The FCBFincreased the best accuracy <strong>on</strong> the peripheral feature to 59% while the ANOVA selecti<strong>on</strong>increased the accuracy to 56% <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG features. The analysis of the classificati<strong>on</strong> accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g>EEG <strong>and</strong> peripheral features computed <strong>on</strong> different durati<strong>on</strong>s dem<strong>on</strong>strated that the EEG featuresare more robust to a decrease in durati<strong>on</strong> than the peripheral features, which c<strong>on</strong>firms theimportance of EEG features <str<strong>on</strong>g>for</str<strong>on</strong>g> short term emoti<strong>on</strong> assessment.From the analysis of the game-over events, distinct patterns of GSR <strong>and</strong> HR activity <str<strong>on</strong>g>for</str<strong>on</strong>g> themedium <strong>and</strong> hard c<strong>on</strong>diti<strong>on</strong>s were found in the 5 sec<strong>on</strong>ds following the event. The distinctpatterns were supposedly due to the differences in arousal <strong>and</strong> pleasantness relative to the gameoverevent <strong>and</strong> to the starting of a new game. Nevertheless, those distinct peripheral patternssuggest that peripheral signals recorded after events occurring during the game could be used todetermine the state of the player.142


<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> of emoti<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> computer gamesFuture work will focus <strong>on</strong> the improvement of the detecti<strong>on</strong> accuracy. Fusi<strong>on</strong> of physiologicalin<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> with other modalities such as facial expressi<strong>on</strong>s <strong>and</strong> speech would certainly improvethe accuracy. Including game in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> such as the evoluti<strong>on</strong> of the score can also help tobetter detect the three states. Another questi<strong>on</strong> of interest is to determine the number of classes tobe detected. Since boredom <strong>and</strong> anxiety are detected with higher c<strong>on</strong>fidence than engagement itmight be enough to use those two classes <str<strong>on</strong>g>for</str<strong>on</strong>g> adaptati<strong>on</strong> to the game difficulty. Moreover, fromthe observati<strong>on</strong> of Figure 7.1, <strong>on</strong>e can c<strong>on</strong>clude that it is more interesting to adapt the difficultyof the game solely based <strong>on</strong> the increase of competence because it leads to a str<strong>on</strong>ger change ofstate in the flow chart (Figure 7.1) <strong>and</strong> stimulates learning. In this case <strong>on</strong>ly the detecti<strong>on</strong> ofboredom is of importance to modulate difficulty. This also implies to more clearly define whatare the relati<strong>on</strong>s between emoti<strong>on</strong>s, competence <strong>and</strong> learning. A future study would be toimplement an adaptive Tetris game <strong>and</strong> verify that it is more fun <strong>and</strong> enjoyable than the st<strong>and</strong>ard<strong>on</strong>e. Finally, analysis of physiological signals <str<strong>on</strong>g>for</str<strong>on</strong>g> different types of games is also required to seeif the results of this study can be extended to other games.143


C<strong>on</strong>clusi<strong>on</strong>sChapter 8C<strong>on</strong>clusi<strong>on</strong>s8.1 OutcomesThis thesis c<strong>on</strong>centrated <strong>on</strong> the study of physiological signals in the c<strong>on</strong>text of affectivecomputing. It aimed at dem<strong>on</strong>strating <strong>and</strong> comparing the usefulness of two categories ofphysiological signals <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment: those that reflect the activity of the central nervoussystem (EEG’s) <strong>and</strong> those reflecting the peripheral nervous system activity (GSR, temperature,BVP, HR <strong>and</strong> respirati<strong>on</strong>).Chapter 2 presented a state of the art of the topics related to emoti<strong>on</strong> assessment fromphysiological signals. Firstly, the most well-known representati<strong>on</strong>s <strong>and</strong> models of emoti<strong>on</strong>s weregiven together with their implicati<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment studies. The valence-arousal spacewas used as the representati<strong>on</strong> of emoti<strong>on</strong>s throughout this thesis since it was c<strong>on</strong>sidered as beingthe most general, flexible <strong>and</strong> less dependent <strong>on</strong> the applicati<strong>on</strong>. The role of c<strong>on</strong>text <strong>and</strong> themultimodal aspects of emoti<strong>on</strong>s were also emphasized from the analysis of the described modelsof emoti<strong>on</strong>s. Sec<strong>on</strong>dly, the physiological processes related to emoti<strong>on</strong>s were described <str<strong>on</strong>g>for</str<strong>on</strong>g> boththe peripheral <strong>and</strong> central activities. A n<strong>on</strong> exhaustive list of the devices usable to record thoseactivities was given <strong>and</strong> the signals features that are known to be related to emoti<strong>on</strong>al processeswere discussed. Thirdly, several studies c<strong>on</strong>cerning the assessment of emoti<strong>on</strong>s fromphysiological signals were reviewed. Six criteri<strong>on</strong>s were proposed to organize, evaluate <strong>and</strong>compare those studies as well as to discuss important aspects of physiological emoti<strong>on</strong>assessment.Chapter 3 started by a descripti<strong>on</strong> of the material used to m<strong>on</strong>itor the physiological activity ofseveral participants during emoti<strong>on</strong>al experiences. While this part is important <str<strong>on</strong>g>for</str<strong>on</strong>g> thereproducibility of the results, it also provides a guide to help pers<strong>on</strong>s that are not used to this typeof apparatus <str<strong>on</strong>g>for</str<strong>on</strong>g> setting up an acquisiti<strong>on</strong> process. The algorithms to pre-process the signals <strong>and</strong>extract the features that characterize the physiological activity were then presented. For EEG’s,features based <strong>on</strong> the energy of the signals were computed according to the literature <strong>and</strong> a new afeature set based <strong>on</strong> the MI between pairs of electrodes was proposed. An algorithm to computeHR from a BVP signal based <strong>on</strong> the detecti<strong>on</strong> of the foot of the systolic upstroke was detailed.The proposed peripheral features were explained <strong>and</strong> discussed according to the literature. Thischapter c<strong>on</strong>cluded <strong>on</strong> the ethical aspects that should be (<strong>and</strong> have been) c<strong>on</strong>sidered whenacquiring physiological signals.Chapter 4 presented the supervised learning methods used to assess emoti<strong>on</strong>s from the extractedphysiological features. It stressed the importance of the ground-truth definiti<strong>on</strong> <strong>and</strong> provideddifferent possibilities to determine the (supposedly) true emoti<strong>on</strong>al state elicited by an event. The145


Chapter 8accuracy <strong>and</strong> the c<strong>on</strong>fusi<strong>on</strong> matrix obtained <strong>on</strong> test data, by using the defined cross-validati<strong>on</strong>strategies, were chosen to measure the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of the emoti<strong>on</strong> assessment. The classifiers(Naïve-Bayes, QDA, LDA, SVM <strong>and</strong> RVM) used to find models that map the physiologicalresp<strong>on</strong>ses to an estimated emoti<strong>on</strong>al state were given. Since some of the extracted feature setswere of high dimensi<strong>on</strong>ality, filter (ANOVA, Fisher based <strong>and</strong> FCBF) <strong>and</strong> wrapper (SFFS)feature selecti<strong>on</strong> algorithms were detailed to later select the features of interest. The last part ofChapter 4 presented the in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> fusi<strong>on</strong> methods. These methods were used in the nextchapters to investigate the usefulness of fusi<strong>on</strong> between the peripheral <strong>and</strong> central in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>.Chapter 5, 6 <strong>and</strong> 7 applied the previously described methods <strong>on</strong> physiological <strong>and</strong> emoti<strong>on</strong>al dataacquired in different elicitati<strong>on</strong> c<strong>on</strong>texts (elicitati<strong>on</strong> method, emoti<strong>on</strong>s elicited <strong>and</strong> durati<strong>on</strong> oftrial). In Chapter 5 the emoti<strong>on</strong>s were elicited by using images from the IAPS <strong>and</strong> the classifierswere independently trained to recover valence or arousal classes. In Chapter 6, the participantshad to self-generate emoti<strong>on</strong>s by remembering past emoti<strong>on</strong>al episodes of their life bel<strong>on</strong>ging tothree areas of the valence-arousal space. Chapter 7 described a gaming protocol that aimed attesting emoti<strong>on</strong> assessment in a c<strong>on</strong>text that is closer to HCI applicati<strong>on</strong>s. The main c<strong>on</strong>clusi<strong>on</strong>s<strong>and</strong> outcomes drawn from those chapters are given below.One of the main outcomes of this thesis is the c<strong>on</strong>clusi<strong>on</strong> that EEG’s are useful <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>assessment. For all the protocols <strong>and</strong> classes <str<strong>on</strong>g>for</str<strong>on</strong>g>mulati<strong>on</strong>s, the average classificati<strong>on</strong> accuracywas higher than the r<strong>and</strong>om level. Moreover, compared to the peripheral features, theclassificati<strong>on</strong> based <strong>on</strong> EEG features generally led a higher accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> the assessment of thevalence dimensi<strong>on</strong> of emoti<strong>on</strong>s. The experiments per<str<strong>on</strong>g>for</str<strong>on</strong>g>med in chapter 6 <strong>and</strong> 7 also dem<strong>on</strong>stratedthat the EEG modality was more adequate than the peripheral modality to assess emoti<strong>on</strong>s <strong>on</strong> ashort time scale.Fusi<strong>on</strong> of peripheral <strong>and</strong> EEG features was shown to be effective as it increased the classificati<strong>on</strong>accuracy, especially <str<strong>on</strong>g>for</str<strong>on</strong>g> the data presented in Chapter 6 <strong>and</strong> 7. The fusi<strong>on</strong> of the classifiersoutputs was always more effective than simple c<strong>on</strong>catenati<strong>on</strong> of the feature vectors. The MIfeature set proposed in Chapter 3, was less effective than the energy feature set. However whenthe two feature sets were fused, an increase of the accuracy was observed showing the interest ofthe MI features <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment.Applying the feature selecti<strong>on</strong> algorithms to reduce the dimensi<strong>on</strong>ality of the extracted featuresled to a str<strong>on</strong>g decrease of the size of the feature sets with <strong>on</strong>ly a reas<strong>on</strong>able decrease (Chapter 6)or increase (Chapter 6 <strong>and</strong> 7) of the accuracy depending <strong>on</strong> the employed classifiers <strong>and</strong> datasets.This is of interest to improve the computati<strong>on</strong>al speed of the classificati<strong>on</strong> algorithms. Moreover,these algorithms assisted in finding the EEG <strong>and</strong> peripheral features that are useful <str<strong>on</strong>g>for</str<strong>on</strong>g> interparticipantclassificati<strong>on</strong> as detailed in Chapter 7.146


C<strong>on</strong>clusi<strong>on</strong>sC<strong>on</strong>cerning the classificati<strong>on</strong> algorithms, n<strong>on</strong>e of the classifiers systematically per<str<strong>on</strong>g>for</str<strong>on</strong>g>med betterthan the others. However, the analysis c<strong>on</strong>ducted in Chapter 6 c<strong>on</strong>firmed the interest of the SVMclassifiers <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> of emoti<strong>on</strong>s in high dimensi<strong>on</strong>al feature spaces.An important outcome of this thesis is the producti<strong>on</strong> of three emoti<strong>on</strong>al databases c<strong>on</strong>taining theperipheral <strong>and</strong> EEG signals of several participants, acquired in different emoti<strong>on</strong>al elicitati<strong>on</strong>c<strong>on</strong>texts. While those databases are not publicly available because of ethical aspects, it remainsthat further studies can be per<str<strong>on</strong>g>for</str<strong>on</strong>g>med <strong>on</strong> this data within the University of Geneva.This thesis has shown the usability of EEG <strong>and</strong> peripheral signals <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment <strong>and</strong>there<str<strong>on</strong>g>for</str<strong>on</strong>g>e encourages their use <str<strong>on</strong>g>for</str<strong>on</strong>g> affective computing. An example of an affective Tetris gamethat adapts its difficulty to the emoti<strong>on</strong>s felt by the user is given in Chapter 7. The resultsobtained from the analysis of physiological signals gathered in Chapter 7 showed that thesesignals are useful <str<strong>on</strong>g>for</str<strong>on</strong>g> the game adaptati<strong>on</strong>. Many other applicati<strong>on</strong>s in areas such as health <strong>and</strong>in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> retrieval can be targeted by this research.8.2 Future prospectsAs dem<strong>on</strong>strated in this thesis, physiological signals can be used to assess emoti<strong>on</strong>s. However,there are still some research issues that have to be investigated in order to improve the accuracyof the assessment <strong>and</strong> apply it to c<strong>on</strong>crete HCI applicati<strong>on</strong>s.<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s can be elicited in several c<strong>on</strong>texts influencing the emoti<strong>on</strong>al expressi<strong>on</strong>. It is thusimportant that the methods developed to assess emoti<strong>on</strong>s take into account those c<strong>on</strong>textualelements. Since new human-computer interfaces will more <strong>and</strong> more involve all of the humansenses it can be interesting to analyze the per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance of emoti<strong>on</strong> assessment algorithmsaccording to the sense used <str<strong>on</strong>g>for</str<strong>on</strong>g> the emoti<strong>on</strong> elicitati<strong>on</strong>. Analyzing the combinati<strong>on</strong>s of suchstimuli can also be of interest. The emoti<strong>on</strong>al model proposed by Ort<strong>on</strong>y [57] is a possibledirecti<strong>on</strong> to follow in order to better determine an emoti<strong>on</strong>al state according to the course ofevents that elicited the emoti<strong>on</strong>. <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment from physiological signals (or from othersources) can then be added to this model to add pers<strong>on</strong>al emoti<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>. The mood of theuser <strong>and</strong> the persistence of the precedent emoti<strong>on</strong>al state are also c<strong>on</strong>text related issues that caninfluence the elicitati<strong>on</strong> of an emoti<strong>on</strong> <strong>and</strong> thus should be taken into account.All the experiments c<strong>on</strong>ducted in this thesis were d<strong>on</strong>e in an “ideal” envir<strong>on</strong>ment where theparticipants were instructed to accomplish given tasks <strong>and</strong> to avoid movements. Switching fromthis type of experiment to the real envir<strong>on</strong>ments gives rise to several issues. Physiological signalsare very sensitive to movements, <str<strong>on</strong>g>for</str<strong>on</strong>g> instance if a user st<strong>and</strong>s up his / her blood pressure willchange. Moreover, the user can be disturbed by external events that are not related to the147


Chapter 8applicati<strong>on</strong> of interest. It is thus very important to develop algorithms that are able to detectsignals changes that are not related to emoti<strong>on</strong>al processes.While this work focused <strong>on</strong> the identificati<strong>on</strong> of emoti<strong>on</strong>al classes defined as areas of thevalence-arousal space, going further toward the identificati<strong>on</strong> of a point in this space is of highinterest to determine emoti<strong>on</strong>al states with higher precisi<strong>on</strong>. This could be useful to infer theintensity of the emoti<strong>on</strong>, generally defined as the distance of the point to the center of the space.Having enough resoluti<strong>on</strong> in this space is also m<strong>and</strong>atory to map a valence-arousal estimati<strong>on</strong> toa given emoti<strong>on</strong>al label. Assessing the dominance (or c<strong>on</strong>trol) dimensi<strong>on</strong> of emoti<strong>on</strong>s can beuseful to well differentiate emoti<strong>on</strong>al states like fear <strong>and</strong> anger. C<strong>on</strong>tinuous estimati<strong>on</strong>s ofemoti<strong>on</strong>al points in those spaces can be d<strong>on</strong>e by using supervised regressi<strong>on</strong> algorithm. In thatcase, since the valence <strong>and</strong> arousal variables seem to be dependent, the regressi<strong>on</strong> should beper<str<strong>on</strong>g>for</str<strong>on</strong>g>med accordingly. Going <str<strong>on</strong>g>for</str<strong>on</strong>g>ward to other c<strong>on</strong>tinuous representati<strong>on</strong>s of emoti<strong>on</strong>s, like theSEC proposed in Scherer’s [10] model is still an opportunity but requires the evaluati<strong>on</strong> of manyc<strong>on</strong>tinuous variables.As dem<strong>on</strong>strated in this thesis, EEG signals can be used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment. However, a lotof ef<str<strong>on</strong>g>for</str<strong>on</strong>g>t should be put <strong>on</strong> the design of new EEG caps that are less obtrusive to go towardapplicati<strong>on</strong>s. For parts of this thesis, 64 electrodes were employed to m<strong>on</strong>itor brain activity. Thishigh number of electrodes is problematic regarding prices aspects <strong>and</strong> leads to high dimensi<strong>on</strong>alfeature spaces. Developing algorithms that are able to select the electrode positi<strong>on</strong>s of interest <str<strong>on</strong>g>for</str<strong>on</strong>g>emoti<strong>on</strong> assessment is of major importance to solve those issues. A possible soluti<strong>on</strong> to thisproblem could be to use the MI computed between pairs of electrodes (as proposed in Chapter 3)to regroup electrodes that recorded similar in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> <strong>and</strong> choose <strong>on</strong>e of them as therepresentative of the group.Increasing the accuracy obtained from EEG features is of major importance <str<strong>on</strong>g>for</str<strong>on</strong>g> practical use ofthis device. For this purpose, new features should be investigated possibly inspired from the BCIcommunity like features based <strong>on</strong> comm<strong>on</strong> spatial patterns. The MI feature set used with successin this study encourages the investigati<strong>on</strong> of interacti<strong>on</strong>s between brain areas during emoti<strong>on</strong>alprocesses. The synchr<strong>on</strong>izati<strong>on</strong> of brain processes could be used to determine new features. Someof the brain structures involved in emoti<strong>on</strong>al processes lie deep in the brain <strong>and</strong> it is thus difficultto assess their activity from surface EEG signals. Solving the inverse problem (i.e. finding thebrain sources corresp<strong>on</strong>ding to a given EEG) to estimate deep sources <strong>and</strong> using this in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>as new features <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment could be promising.Fusi<strong>on</strong> with other sensors <strong>and</strong> sources of emoti<strong>on</strong>al in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> could lead to improvements ofthe emoti<strong>on</strong> assessment accuracy. Several sensors can be used to acquire signals originating fromthe same sources. For instance, EEG measurements can be coupled with fNIRS measurements to148


C<strong>on</strong>clusi<strong>on</strong>sbetter estimate brain activity. However, since emoti<strong>on</strong>s are multi-modal processes that involveseveral comp<strong>on</strong>ent of the organism it is certainly most valuable to per<str<strong>on</strong>g>for</str<strong>on</strong>g>m the fusi<strong>on</strong> of differentsources of in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>. For instance, this could be achieved by combining facial expressi<strong>on</strong> <strong>and</strong>speech identificati<strong>on</strong> with physiological measurements of emoti<strong>on</strong>s. Those fusi<strong>on</strong>s wouldnecessitate more studies, especially c<strong>on</strong>cerning time aspects. The time resoluti<strong>on</strong> of differentsensors is not the same <strong>and</strong> the different comp<strong>on</strong>ents involved in emoti<strong>on</strong>al processes do not havethe same reacti<strong>on</strong> time.Most of the studies c<strong>on</strong>cerning emoti<strong>on</strong> assessment from physiological signals (including thisthesis) are d<strong>on</strong>e <strong>on</strong> emoti<strong>on</strong>al data that are not available to the whole research community. As ac<strong>on</strong>sequence it is difficult to compare the methods used <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong> assessment since theirper<str<strong>on</strong>g>for</str<strong>on</strong>g>mances str<strong>on</strong>gly depend <strong>on</strong> the protocol used <str<strong>on</strong>g>for</str<strong>on</strong>g> data acquisiti<strong>on</strong>. There is thus animportant need <str<strong>on</strong>g>for</str<strong>on</strong>g> databases that are freely accessible. Such databases should ideally bemultimodal, include c<strong>on</strong>textual in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> <strong>and</strong> meet the c<strong>on</strong>straints imposed by the law / ethicalrules. A freely available multimodal database 10 of emoti<strong>on</strong>ally driven brain <strong>and</strong> peripheral signalswas c<strong>on</strong>structed in collaborati<strong>on</strong> with partners of the European Network of excellence Similar.Un<str<strong>on</strong>g>for</str<strong>on</strong>g>tunately, this data was not analyzed in this thesis but we str<strong>on</strong>gly encourage the use of thisdatabase <str<strong>on</strong>g>for</str<strong>on</strong>g> further research <strong>on</strong> the topic. A similar ef<str<strong>on</strong>g>for</str<strong>on</strong>g>t is now underway in the c<strong>on</strong>text of theEU project Petamedia.Taken together, the above suggesti<strong>on</strong>s should lead to the development of a robust emoti<strong>on</strong>assessment system. Once such a system is developed the next step would be to determine how themachine should adapt to the user’s emoti<strong>on</strong>al state. Some propositi<strong>on</strong>s were given in Chapter 7<str<strong>on</strong>g>for</str<strong>on</strong>g> computer games but this strategy is highly dependent <strong>on</strong> the gaming applicati<strong>on</strong>. Finally, theinvestigati<strong>on</strong> of how the user perceives the complete system (i.e. emoti<strong>on</strong> assessment <strong>and</strong>adaptati<strong>on</strong>) is m<strong>and</strong>atory to c<strong>on</strong>trol how it would be received by the general public.10 available at http://enterface.tel.fer.hr/index.php?frame=results (retrieved <strong>on</strong> 19 May 2009)149


C<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>mAppendix A C<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m151


Appendix A152


C<strong>on</strong>sent <str<strong>on</strong>g>for</str<strong>on</strong>g>m153


Neighborhood table <str<strong>on</strong>g>for</str<strong>on</strong>g> the Laplacian filterAppendix B Neighborhood table <str<strong>on</strong>g>for</str<strong>on</strong>g> the Laplacian filterThe neighborhood of an electrode was defined as proposed in [164]. The following table gives<str<strong>on</strong>g>for</str<strong>on</strong>g> each electrode the list of the associated neighbors.Electrode Neighbors electrodes Electrode Neighbors electrodesFp1 F7, F5, AF7, AFz, Fpz Fpz Fp1, AFz, Fp2AF7 Fp1, F5, F3, AF3, AFz Fp2 Fpz, AFz, AF8, F6, F8AF3 AFz, AF7, F3, F1, Fz, AF4 AF8 Fp2, AFz, AF4, F4, F6F1 F3, FC3, FC1, Fz, AF3 AF4 AF8, AFz, AF3, Fz, F2, F4F3 F5, FC5, FC3, F1, AF3, AF7 AFz Fpz, Fp1, AF7, AF3, AF4, AF8, Fp2F5 F7, FT7, FC5, F3, AF7, Fp1 Fz AF3, F1, FC1, FCz, FC2, F2, AF4F7 FT7, F5, Fp1 F2 AF4, Fz, FC2, FC4, F4FT7 T7, FC5, F5, F7 F4 F6, AF8, AF4, F2, FC4, FC6FC5 FT7, T7, C5, FC3, F3, F5 F6 F8, Fp2, AF8, F4, FC6, FT8FC3 C5, C3, C1, FC1, F1, F3, FC5 F8 Fp2, F6, FT8FC1 F1, FC3, C1, FCz, Fz FT8 F8, F6, FC6, T8C1 FCz, FC1, FC3, C3, CP1, Cz FC6 FT8, F6, F4, FC4, C6, T8C3 C1, FC3, C5, CP3, CP1 FC4 FC6, F4, F2, FC2, C2, C4, C6C5 CP5, CP3, C3, FC3, FC5, T7 FC2 F2, Fz, FCz, C2, FC4T7 TP7, CP5, C5, FC5, FT7 FCz Fz, FC1, C1, Cz, C2, FC2TP7 P9, P7, CP5, T7 Cz FCz, C1, CP1, CPz, CP2, C2CP5 TP7, P7, P5, CP3, C5, T7 C2 FC4, FC2, FCz, Cz, CP2, C4CP3 CP5, P5, P3, CP1, C3, C5 C4 C6, FC4, C2, CP2, CP4CP1 Cz, C1, C3, CP3, P3, Pz, CPz C6 T8, FC6, FC4, C4, CP4, CP6P1 POz, P2, Pz, P3, P5, PO3 T8 FT8, FC6, C6, CP6, TP8P3 P5, P1, Pz, CP1, CP3 TP8 T8, CP6, P8, P10P5 PO3, P1, P3, CP3, CP5, P7 CP6 TP8, T8, C6, CP4, P6, P8P7 PO7, PO3, P5, CP5, TP7, P9 CP4 CP6, C6, C4, CP2, P4, P6P9 O1, PO7, P7, TP7 CP2 C4, C2, Cz, CPz, Pz, P4, CP4PO7 Iz, Oz, PO3, P7, P9, O1 P2 PO4, P6, P4, Pz, P1, POzPO3 Oz, POz, P1, P5, P7, PO7 P4 P6, CP4, CP2, Pz, P2O1 Iz, PO7, P9 P6 P8, CP6, CP4, P4, P2, PO4Iz O2, PO8, Oz, PO7, O1 P8 TP8, CP6, P6, PO4, PO8, P10Oz Iz, PO8, PO4, POz, PO3, PO7 P10 TP8, P8, PO8, O2POz Oz, PO4, P2, P1, PO3 PO8 O2, P10, P8, PO4, Oz, IzPz P2, P4, CP2, CPz, CP1, P3, P1 PO4 P8, P6, P2, POz, Oz, PO8CPz Cz, CP1, Pz, CP2 O2 P10, PO8, Iz155


List of IAPS images usedAppendix C List of IAPS images usedListe of IAPS images <str<strong>on</strong>g>for</str<strong>on</strong>g> the arousal experimentIAPS imagenumberAssociatedclassIAPS imagenumberAssociatedclassIAPS imagenumberAssociatedclass1050 High arousal 4659 High arousal 7050 Low arousal1201 High arousal 4800 High arousal 7060 Low arousal1300 High arousal 5130 Low arousal 7080 Low arousal1310 High arousal 5390 Low arousal 7090 Low arousal1931 High arousal 5470 High arousal 7100 Low arousal2190 Low arousal 5500 Low arousal 7110 Low arousal2381 Low arousal 5510 Low arousal 7140 Low arousal2440 Low arousal 5520 Low arousal 7150 Low arousal2480 Low arousal 5530 Low arousal 7175 Low arousal2570 Low arousal 5621 High arousal 7185 Low arousal2580 Low arousal 5623 High arousal 7187 Low arousal2620 Low arousal 5626 High arousal 7205 Low arousal2661 High arousal 5700 High arousal 7217 Low arousal2691 High arousal 5731 Low arousal 7224 Low arousal2840 Low arousal 5740 Low arousal 7233 Low arousal2850 Low arousal 5910 High arousal 7234 Low arousal2870 Low arousal 5920 High arousal 7235 Low arousal2880 Low arousal 5940 High arousal 7380 High arousal2890 Low arousal 5950 High arousal 7490 Low arousal3030 High arousal 6570 High arousal 7491 Low arousal3053 High arousal 7000 Low arousal 7640 High arousal3071 High arousal 7004 Low arousal 7950 Low arousal3080 High arousal 7006 Low arousal 8030 High arousal3150 High arousal 7010 Low arousal 8080 High arousal3170 High arousal 7020 Low arousal 8160 High arousal3261 High arousal 7025 Low arousal 8161 High arousal4220 High arousal 7031 Low arousal 8170 High arousal4290 High arousal 7035 Low arousal 8200 High arousal4658 High arousal 7040 Low arousal 8300 High arousal8400 High arousal 9050 High arousal 9600 High arousal8490 High arousal 9360 Low arousal 9622 High arousal8500 High arousal 9405 High arousal 9810 High arousal8501 High arousal 9570 High arousal9040 High arousal 9571 High arousal157


Appendix CListe of IAPS images <str<strong>on</strong>g>for</str<strong>on</strong>g> the valence experimentIAPS imagenumberAssociatedclassIAPS imagenumberAssociatedclassIAPS imagenumberAssociatedclass1710 Positive 5621 Positive 8501 Positive1811 Positive 5623 Positive 8502 Positive2053 Negative 5629 Positive 8503 Positive2160 Positive 5660 Positive 8531 Positive2710 Negative 5700 Positive 8540 Positive2800 Negative 5910 Positive 9006 Negative3051 Negative 6212 Negative 9050 Negative3060 Negative 6230 Negative 9140 Negative3063 Negative 6300 Negative 9181 Negative3064 Negative 6360 Negative 9252 Negative3100 Negative 6830 Negative 9253 Negative3110 Negative 6831 Negative 9300 Negative3120 Negative 7230 Positive 9340 Negative3130 Negative 7260 Positive 9400 Negative3180 Negative 7270 Positive 9405 Negative3220 Negative 7330 Positive 9421 Negative3230 Negative 7380 Negative 9433 Negative3400 Negative 7502 Positive 9500 Negative4220 Positive 8030 Positive 9520 Negative4599 Positive 8034 Positive 9560 Negative4607 Positive 8080 Positive 9570 Negative4608 Positive 8090 Positive 9571 Negative4610 Positive 8170 Positive 9600 Negative4640 Positive 8180 Positive 9611 Negative4641 Positive 8190 Positive 9620 Negative4660 Positive 8200 Positive 9630 Negative4680 Positive 8210 Positive 9800 Negative5260 Positive 8230 Negative 9810 Negative5270 Positive 8350 Positive 9910 Negative5450 Positive 8370 Positive 9911 Negative5460 Positive 8380 Positive 9920 Negative5470 Positive 8400 Positive 9921 Negative5480 Positive 8420 Positive5600 Positive 8496 PositiveThe images that are comm<strong>on</strong> to the arousal <strong>and</strong> valence experiment are: 4220, 5470, 5621, 5623,5700, 5910, 7380, 8030, 8080, 8170, 8200, 8400, 8501, 9050, 9405, 9570, 9571, 9600, 9810.158


Questi<strong>on</strong>naire results <str<strong>on</strong>g>for</str<strong>on</strong>g> the game protocolAppendix D Questi<strong>on</strong>naire results <str<strong>on</strong>g>for</str<strong>on</strong>g> the game protocolThe following table gives the list of the 30 statements that the participants had to evaluate usingLikert scales. The participants also had to evaluate their emoti<strong>on</strong> using the valence, arousal <strong>and</strong>dominance SAM scales.The weights higher than 0.7 are highlighted in gray.Statement 1 st comp<strong>on</strong>ent 2 nd comp<strong>on</strong>entQ1 J'ai apprécié le jeu 0.90 -0.09Q2 J'ai été intéressé(e) 0.90 0.13Q3 J'ai essayé de nouvelles comm<strong>and</strong>es 0.34 -0.03Q4J'ai été motivé(e) à faire le meilleur scorepossible 0.74 -0.17Q5 J'ai eu du plaisir à jouer 0.90 -0.17Q6 J'ai dû m'adapter aux comm<strong>and</strong>es 0.35 0.41Q7 J'ai trouvé la partie difficile -0.12 0.87Q8 J'ai été absorbé(e) par le jeu 0.60 0.56Q9J'ai pris en compte d'avantage d'in<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>qu'auparavant 0.55 -0.16Q10 J'ai dû réfléchir 0.75 0.00Q11 J'ai été stressé(e) 0.06 0.88Q12 M<strong>on</strong> attenti<strong>on</strong> s'est focalisée sur la partie 0.53 0.50Q13 J'ai eu l'impressi<strong>on</strong> que je m'améliorais 0.66 0.24Q14 J'aurais pu obtenir un meilleur score 0.54 0.00Q15 J'ai été amusé(e) 0.86 0.07Q16 Je n'ai pas vu le temps passer 0.50 0.45Q17 J'ai joué au maximum de mes capacités 0.36 0.48Q18 J'ai oublié ce qui se passait autour de moi 0.32 0.44Q19 J'ai trouvé la partie trop facile 0.11 -0.84Q20 J'ai ressenti de l'ennui -0.46 -0.21Q21 J'ai senti que j'avais le c<strong>on</strong>trôle 0.40 -0.78Q22 J'ai trouvé la partie plaisante 0.88 -0.10Q23 J'ai été calme 0.19 -0.67Q24 J'aurais vol<strong>on</strong>tiers c<strong>on</strong>tinué à jouer 0.86 -0.11Q25 J'ai été excité(e) 0.31 0.70Q26 J'ai su clairement ce que je devais faire 0.35 -0.49Q27 J'ai pu diriger les pièces comme je le voulais 0.34 -0.82Q28 J'ai été c<strong>on</strong>centré(e) totalement sur le jeu 0.60 0.46Q29 J'ai été mis(e) sous pressi<strong>on</strong> -0.02 0.81Q30 J'ai pensé à d'autres choses qu'au jeu -0.12 -0.45Q31 valence (SAM) 0.73 0.01Q32 arousal (SAM) 0.17 0.77Q33 dominance (SAM) 0.37 -0.59159


Publicati<strong>on</strong>sAppendix E Publicati<strong>on</strong>sRefereed journals, special issuesG. Chanel, J.J.M. Kierkels, M. Soleymani, T. Pun, "Short-term emoti<strong>on</strong> assessment in a recallparadigm", Internati<strong>on</strong>al Journal of Human-Computer Studies. Accepted <str<strong>on</strong>g>for</str<strong>on</strong>g> publicati<strong>on</strong>, March2009, DOI:10.1016/j.ijhcs.2009.03.005M. Soleymani, G. Chanel, J.J.M. Kierkels, T. Pun, "<str<strong>on</strong>g>Affective</str<strong>on</strong>g> characterizati<strong>on</strong> of movie scenesbased <strong>on</strong> c<strong>on</strong>tent analysis <strong>and</strong> physiological changes", Internati<strong>on</strong>al Journal <strong>on</strong> Semantic<str<strong>on</strong>g>Computing</str<strong>on</strong>g>; submitted, Feb. 2009.J. Kr<strong>on</strong>egg, G. Chanel, S. Voloshynovskiy, T. Pun, "EEG-based synchr<strong>on</strong>ized brain-computerinterfaces: a model <str<strong>on</strong>g>for</str<strong>on</strong>g> optimizing the number of mental tasks", IEEE Transacti<strong>on</strong>s <strong>on</strong> NeuralSystems <strong>and</strong> Rehabilitati<strong>on</strong> Engineering, Vol. 15, No. 1, March 2007, pp. 50-58.T. Pun, T. I. Alecu, G. Chanel, J. Kr<strong>on</strong>egg, S. Voloshynovskiy, "<strong>Brain</strong>-computer interacti<strong>on</strong>research at the Computer Visi<strong>on</strong> <strong>and</strong> Multimedia Laboratory, University of Geneva", IEEETransacti<strong>on</strong>s Neural Systems <strong>and</strong> Rehabilitati<strong>on</strong> Engineering, Special Issue <strong>on</strong> <strong>Brain</strong>-ComputerInteracti<strong>on</strong>, T. M. Vaughan <strong>and</strong> J. R. Wolpaw, Eds., Vol. 14, No. 2, June 2006, pp. 210-213.Internati<strong>on</strong>al c<strong>on</strong>ferences with refereed full length articlesG. Chanel, C. Rebetez, M. Betrancourt, T. Pun, "Boredom, engagement <strong>and</strong> anxiety as indicators<str<strong>on</strong>g>for</str<strong>on</strong>g> adaptati<strong>on</strong> to difficulty in games", 11th MindTrek C<strong>on</strong>ference, MindTrek 2008: Entertainment<strong>and</strong> Media in the Ubiquitous Era, Tampere, Finl<strong>and</strong>, October 7-9, 2008.G. Chanel, K. Ansari-Asl, T. Pun, "Valence-arousal evaluati<strong>on</strong> using physiological signals in anemoti<strong>on</strong> recall paradigm", 2007 IEEE SMC, Internati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> Systems, Man <strong>and</strong>Cybernetics, Smart cooperative systems <strong>and</strong> cybernetics: advancing knowledge <strong>and</strong> security <str<strong>on</strong>g>for</str<strong>on</strong>g>humanity, M<strong>on</strong>treal, Canada, Oct. 7-10, 2007.G. Chanel. J. Kr<strong>on</strong>egg, D. Gr<strong>and</strong>jean, I. Alecu, T. Pun, "Pattern recogniti<strong>on</strong> in peripheral <strong>and</strong>central signaling", Workshop <strong>on</strong> Multimodal synchr<strong>on</strong>izati<strong>on</strong> in affective expressi<strong>on</strong>s, HumaineEuropean NOE 3rd Summer School <strong>and</strong> <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Sciences NCCR, Genova, Italy, September 22-24, 2006. Invited l<strong>on</strong>g presentati<strong>on</strong>.G. Chanel, J. Kr<strong>on</strong>egg, D. Gr<strong>and</strong>jean, T. Pun, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment: Arousal evaluati<strong>on</strong> usingEEG's <strong>and</strong> peripheral physiological signals", Proceedings of the Internati<strong>on</strong>al Workshop <strong>on</strong>Multimedia C<strong>on</strong>tent Representati<strong>on</strong>, Classificati<strong>on</strong> <strong>and</strong> Security (MRCS), Special Sessi<strong>on</strong>:Multimodal Signal Processing, Istanbul, Turkey, Sept. 11-13, 2006, B. Gunsel, A. K. Jain, A. M.Tekalp, B. Sankur, Eds., Lecture Notes in Computer Science, Vol. 4105, Springer, 530-537.161


Appendix EM. Soleymani, J.J.M. Kierkels, G. Chanel, T. Pun, "A Bayesian framework <str<strong>on</strong>g>for</str<strong>on</strong>g> video affectiverepresentati<strong>on</strong>", ACII 2009, Internati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> <str<strong>on</strong>g>Affective</str<strong>on</strong>g> <str<strong>on</strong>g>Computing</str<strong>on</strong>g> <strong>and</strong> IntelligentInteracti<strong>on</strong>, Amsterdam, The Netherl<strong>and</strong>s, Sept. 10-12, 2009. Subm. April. 2009.M. Soleymani, G. Chanel, J.J.M. Kierkels, T. Pun, "<str<strong>on</strong>g>Affective</str<strong>on</strong>g> characterizati<strong>on</strong> of movie scenesbased <strong>on</strong> multimedia c<strong>on</strong>tent analysis <strong>and</strong> user's physiological emoti<strong>on</strong>al resp<strong>on</strong>ses", ISM2008,IEEE Internati<strong>on</strong>al Symposium <strong>on</strong> Multimedia, Berkeley, Cali<str<strong>on</strong>g>for</str<strong>on</strong>g>nia, USA, December 15-17,2008, 228-235.M. Soleymani, G. Chanel, J.J.M. Kierkels, T. Pun, "<str<strong>on</strong>g>Affective</str<strong>on</strong>g> ranking of movie scenes usingphysiological signals <strong>and</strong> c<strong>on</strong>tent analysis", ACM Multimedia 2008, Workshop <strong>on</strong> Many Faces ofMultimedia Semantics (MS'08), Vancouver, BC, Canada, Oct. 27-31, 2008, pp. 32-39.M. Soleymani, G. Chanel, J.J.M. Kierkels, T. Pun, "Valence-arousal representati<strong>on</strong> of moviescenes based <strong>on</strong> multimedia c<strong>on</strong>tent analysis <strong>and</strong> user's physiological emoti<strong>on</strong>al resp<strong>on</strong>ses",MLMI 2008, 5th Joint Workshop <strong>on</strong> Machine Learning <strong>and</strong> Multimodal Interacti<strong>on</strong>, Utrecht, TheNetherl<strong>and</strong>s, 8-10 Sept. 2008, (extended abstract <strong>and</strong> poster).K. Ansari-Asl, G. Chanel, T. Pun, "A channel selecti<strong>on</strong> method <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG classificati<strong>on</strong> in emoti<strong>on</strong>assessment based <strong>on</strong> synchr<strong>on</strong>izati<strong>on</strong> likelihood", Eusipco 2007, 15th European SignalProcessing C<strong>on</strong>ference, Poznan, Pol<strong>and</strong>, Sept. 3-7, 2007.A. Benoit, L. B<strong>on</strong>naud, A. Caplier, P. Ngo, L. Laws<strong>on</strong>, D. Trevisan, V. Levacik, C. Mancas, <strong>and</strong>G. Chanel, "Multimodal Focus Attenti<strong>on</strong> <strong>and</strong> Stress Detecti<strong>on</strong> <strong>and</strong> Feedback in an AugmentedDriver Simulator", 3rd IFIP C<strong>on</strong>ference <strong>on</strong> Artificial Intelligence Applicati<strong>on</strong>s <strong>and</strong> Innovati<strong>on</strong>s(AIAI), Athens, June 7-9, 2006.A. Savran, K. Ciftci, G. Chanel, J. Cruz Mota, L. H<strong>on</strong>g Viet, B. Sankur, L. Akarun, A. Caplier<strong>and</strong> M. Rombaut, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> Detecti<strong>on</strong> in the Loop from <strong>Brain</strong> Signals <strong>and</strong> Facial Images",Proceedings of eNTERFACE 2006 Workshop, Dubrovnik, Croatia, July - August 2006.J. Kr<strong>on</strong>egg, T. Alecu, G. Chanel, S. Voloshynovskiy, T. Pun, "Analyse des mesures de débit pourinterfaces cerveau-ordinateur", TAIMA'05, Traitement et Analyse de l'In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> : Méthodes etApplicati<strong>on</strong>s, Hammamet, Tunisie, 26 sep - 1 oct, 2005.Other c<strong>on</strong>ferencesG. Chanel, J.J.M. Kierkels, M. Soleymani, D. Gr<strong>and</strong>jean, T. Pun, "Short-term emoti<strong>on</strong>assessment in a recall paradigm", Joint (IM)2-Interactive Multimodal In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> Management<strong>and</strong> <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Sciences NCCRs Workshop, Riederalp, Switzerl<strong>and</strong>, Sept. 1-3, 2008.G. Chanel, S. Delplanque, "Olfactory-elicited emoti<strong>on</strong>s assessment through psychophysiologicalsignals: statistical <strong>and</strong> classificati<strong>on</strong> approaches'', Joint (IM)2-Interactive Multimodal In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>Management <strong>and</strong> <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Sciences NCCRs Workshop, Riederalp, Switzerl<strong>and</strong>, Sept. 1-3, 2008.162


Publicati<strong>on</strong>sG. Chanel, C. Rebetez, M. Betrancourt, T. Pun, "Boredom, engagement <strong>and</strong> anxiety as indicators<str<strong>on</strong>g>for</str<strong>on</strong>g> adaptati<strong>on</strong> to difficulty in games'', Joint (IM)2-Interactive Multimodal In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>Management <strong>and</strong> <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Sciences NCCRs Workshop, Riederalp, Switzerl<strong>and</strong>, Sept. 1-3, 2008.G. Chanel, K. Ansari-Asl, T. Pun, "From thoughts to emoti<strong>on</strong>s: emoti<strong>on</strong>al state assessment usingphysiological recordings", HUMABIO EU Project Workshop, Securing infrastructures <strong>and</strong>enhancing safety in critical operati<strong>on</strong>s; Humabio physiological <strong>and</strong> behavioural biometrics <str<strong>on</strong>g>for</str<strong>on</strong>g>unobtrusive authenticati<strong>on</strong> <strong>and</strong> m<strong>on</strong>itoring, Basel, Switzerl<strong>and</strong>, February 2, 2007.M. Soleymani, G. Chanel, J.J.M. Kierkels, T. Pun, "Valence-arousal representati<strong>on</strong> of moviescenes based <strong>on</strong> multimedia c<strong>on</strong>tent analysis <strong>and</strong> user's physiological emoti<strong>on</strong>al resp<strong>on</strong>ses", 2008PetaMedia Workshop <strong>on</strong> Implicit, Human-Centered Tagging (HCT), Queen Mary UniversityL<strong>on</strong>d<strong>on</strong>, L<strong>on</strong>d<strong>on</strong>, UK, Sept. 5, 2008.M. Soleymani, J.J.M. Kierkels, G. Chanel, E. Bruno, S. March<strong>and</strong>-Maillet, T. Pun, "Estimatingemoti<strong>on</strong>s <strong>and</strong> tracking interest during movie watching based <strong>on</strong> multimedia c<strong>on</strong>tent <strong>and</strong>physiological resp<strong>on</strong>ses'', Joint (IM)2-Interactive Multimodal In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> Management <strong>and</strong><str<strong>on</strong>g>Affective</str<strong>on</strong>g> Sciences NCCRs Workshop, Riederalp, Switzerl<strong>and</strong>, Sept. 1-3, 2008.K. Ansari-Asl, G. Chanel, T. Pun, "A Channel selecti<strong>on</strong> method <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG classificati<strong>on</strong> inemoti<strong>on</strong> assessment based <strong>on</strong> synchr<strong>on</strong>izati<strong>on</strong> likelihood", Similar NOE Workshop, University ofMagdeburg, Germany, June 4-5, 2007.J. Kr<strong>on</strong>egg, G. Chanel, S. Voloshynovskiy, T. Pun, "In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>-transfer rate based per<str<strong>on</strong>g>for</str<strong>on</strong>g>manceoptimizati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> brain-computer interfaces", Similar NOE Workshop, Herakli<strong>on</strong>, Greece, June 8-9, 2006.T. Pun, T. I. Alecu, G. Chanel, J. Kr<strong>on</strong>egg, S. Voloshynovskiy, "Research in <strong>Brain</strong>-computerinteracti<strong>on</strong>, Multimodal Interacti<strong>on</strong> Group, Computer Visi<strong>on</strong> <strong>and</strong> Multimedia Laboratory,University of Geneva", BCI 2005, <strong>Brain</strong>-Computer Interface Technology: Third Internati<strong>on</strong>alMeeting, Rensselaerville, NY, USA, June 14-19, 2005.163


ReferencesReferences[1] D. Goleman, <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al Intelligence: Why It Can Matter More Than IQ: Bantam, 1997.[2] R. W. Picard, <str<strong>on</strong>g>Affective</str<strong>on</strong>g> computing: The MIT Press, 1997.[3] S. Brave, C. Nass, <strong>and</strong> K. Hutchins<strong>on</strong>, "Computers that care: investigating the effects o<str<strong>on</strong>g>for</str<strong>on</strong>g>ientati<strong>on</strong> of emoti<strong>on</strong> exhibited by an embodied computer agent," Internati<strong>on</strong>al Journalof Human Computer Studies, vol. 62, No. 2, 2005, pp. 161-178.[4] Z. Xu, D. John, <strong>and</strong> A. C. Boucouvalas, "Expressive image generati<strong>on</strong>: Towardsexpressive Internet communicati<strong>on</strong>s," Journal of Visual Languages <strong>and</strong> <str<strong>on</strong>g>Computing</str<strong>on</strong>g>, vol.17, No. 5, Oct 2006, pp. 445-465.[5] P. Ekman, R. W. Levens<strong>on</strong>, <strong>and</strong> W. V. Friesen, "Aut<strong>on</strong>omic Nervous-System ActivityDistinguishes am<strong>on</strong>g <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s," Science, vol. 221, No. 4616, 1983, pp. 1208-1210.[6] P. Rainville, A. Bechara, N. Naqvi, <strong>and</strong> A. R. Damasio, "Basic emoti<strong>on</strong>s are associatedwith distinct patterns of cardiorespiratory activity," Internati<strong>on</strong>al Journal ofPsychophysiology, vol. 61, No. 1, 2006, pp. 5-18.[7] P. J. Lang, M. K. Greenwald, M. M. Bradley, <strong>and</strong> A. O. Hamm, "Looking at pictures:affective, facial, visceral, <strong>and</strong> behavioral reacti<strong>on</strong>s," Psychophysiology, vol. 30, No. 3,May 1993, pp. 261-273.[8] D. A. Norman, The Psychology of Everyday Things. New York: Doubleday / Currency,1990.[9] D. S<strong>and</strong>er, D. Gr<strong>and</strong>jean, <strong>and</strong> K. R. Scherer, "A systems approach to appraisalmechanisms in emoti<strong>on</strong>," Neural Networks, vol. 18, No. 4, May 2005, pp. 317-352.[10] K. R. Scherer, Appraisal c<strong>on</strong>sidered as a process of multi-level sequential checking.Ox<str<strong>on</strong>g>for</str<strong>on</strong>g>d: Ox<str<strong>on</strong>g>for</str<strong>on</strong>g>d University Press, 2001.[11] R. R. Cornelius, The Science of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>. Upper Saddle River, NJ: Prentice-Hall, 1996.[12] R. L. M<strong>and</strong>ryk <strong>and</strong> M. S. Atkins, "A fuzzy physiological approach <str<strong>on</strong>g>for</str<strong>on</strong>g> c<strong>on</strong>tinuouslymodeling emoti<strong>on</strong> during interacti<strong>on</strong> with play technologies," Internati<strong>on</strong>al Journal ofHuman-Computer Studies, vol. 65, No. 4, Apr 2007, pp. 329-347.[13] M. Prensky, "Computer games <strong>and</strong> learning : digital game-based learning," in H<strong>and</strong>bookof computer games studies, J. Raessens <strong>and</strong> J. Goldstein, Eds. Cambridge, MA: The MITPress, 2005.[14] J. M. Carroll, "Learning in Communities, Interdisciplinary Persepectives <strong>on</strong> HumanCentered In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> Technology," in Human-computer interacti<strong>on</strong> series, J. Karat, Ed.L<strong>on</strong>d<strong>on</strong>: Springer, 2009.[15] M. Kankaanranta <strong>and</strong> P. Neittaanmäki, "Design <strong>and</strong> Use of Serious Games," in Intelligentsystems, c<strong>on</strong>trol, <strong>and</strong> automati<strong>on</strong>: science <strong>and</strong> engineering, vol. 37, S. G. Tzafestas, Ed.L<strong>on</strong>d<strong>on</strong>: Springer, 2009.[16] W. Burles<strong>on</strong> <strong>and</strong> R. W. Picard, "Gender-specific approaches to developing emoti<strong>on</strong>allyintelligent learning compani<strong>on</strong>s," IEEE Intelligent Systems, vol. 22, No. 4, Jul-Aug 2007,pp. 62-69.[17] S. H. Fairclough, "Psychophysiological inference <strong>and</strong> physiological computer games,"<strong>Brain</strong>play’07: <strong>Brain</strong>-Computer Interfaces <strong>and</strong> Games, Workshop at the Int. C<strong>on</strong>f. <strong>on</strong>Advances in Computer Entertainment, Salzburg, Austria June 13-15, 2007.[18] B. Cowley, D. Charles, <strong>and</strong> M. Black, "Toward an Underst<strong>and</strong>ing of Flow in VideoGames," ACM Computers in Entertainment vol. 6, No. 2, July 2008.165


References[19] S. K. Sutt<strong>on</strong> <strong>and</strong> R. J. Davids<strong>on</strong>, "Prefr<strong>on</strong>tal brain asymmetry: A biological substrate ofthe behavioral approach <strong>and</strong> inhibiti<strong>on</strong> systems," Psychological Science, vol. 8, No. 3,May 1997, pp. 204-210.[20] R. Cowie, E. Douglas-Cowie, N. Tsapatsoulis, G. Votsis, S. Kollias, W. Fellenz, <strong>and</strong> J. G.Taylor, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> recogniti<strong>on</strong> in human-computer interacti<strong>on</strong>," IEEE Signal ProcessingMagazine, vol. 18, No. 1, Jan 2001, pp. 32-80.[21] S. Helal, W. Mann, H. El-Zabadani, J. King, Y. Kaddoura, <strong>and</strong> E. Jansen, "The GatorTech Smart House: A programmable pervasive space," Computer, vol. 38, No. 3, Mar2005, pp. 50-+.[22] A. Hanjalic, "Adaptive extracti<strong>on</strong> of highlights from a sport video based <strong>on</strong> excitementmodeling," IEEE Transacti<strong>on</strong>s <strong>on</strong> Multimedia, vol. 7, No. 6, Dec 2005, pp. 1114-1122.[23] A. Hanjalic <strong>and</strong> L.-Q. Xu, "<str<strong>on</strong>g>Affective</str<strong>on</strong>g> video c<strong>on</strong>tent representati<strong>on</strong> <strong>and</strong> modeling," IEEETransacti<strong>on</strong>s <strong>on</strong> Multimedia, vol. 7, No. 1, February 2005, pp. 143-154.[24] H. L. Wang <strong>and</strong> L. F. Che<strong>on</strong>g, "<str<strong>on</strong>g>Affective</str<strong>on</strong>g> underst<strong>and</strong>ing in film," Ieee Transacti<strong>on</strong>s <strong>on</strong>Circuits <strong>and</strong> Systems <str<strong>on</strong>g>for</str<strong>on</strong>g> Video Technology, vol. 16, No. 6, Jun 2006, pp. 689-704.[25] C. C. Liu, K. C<strong>on</strong>n, N. Sarkar, <strong>and</strong> W. St<strong>on</strong>e, "Online affect detecti<strong>on</strong> <strong>and</strong> robot behavioradaptati<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> interventi<strong>on</strong> of children with autism," IEEE Transacti<strong>on</strong>s <strong>on</strong> Robotics, vol.24, No. 4, Aug 2008, pp. 883-896.[26] J. R. Wolpaw, N. Birbaumer, W. J. Heetderks, D. J. McFarl<strong>and</strong>, P. H. Peckham, G.Schalk, E. D<strong>on</strong>chin, L. A. Quatrano, C. J. Robins<strong>on</strong>, <strong>and</strong> T. M. Vaughan, "<strong>Brain</strong>computerinterface technology: A review of the first internati<strong>on</strong>al meeting," IEEETransacti<strong>on</strong>s <strong>on</strong> Rehabilitati<strong>on</strong> Engineering, vol. 8, No. 2, Jun 2000, pp. 164-173.[27] T. M. Vaughan, W. J. Heetderks, L. J. Trejo, W. Z. Rymer, M. Weinrich, M. M. Moore,A. Kubler, B. H. Dobkin, N. Birbaumer, E. D<strong>on</strong>chin, E. W. Wolpaw, <strong>and</strong> J. R. Wolpaw,"<strong>Brain</strong>-computer interface technology: A review of the sec<strong>on</strong>d internati<strong>on</strong>al meeting,"IEEE Transacti<strong>on</strong>s <strong>on</strong> Neural Systems <strong>and</strong> Rehabilitati<strong>on</strong> Engineering, vol. 11, No. 2, Jun2003, pp. 94-109.[28] J. Kr<strong>on</strong>egg, "Mesures et optimisati<strong>on</strong> des per<str<strong>on</strong>g>for</str<strong>on</strong>g>mances pour une interface decommunicati<strong>on</strong> neur<strong>on</strong>ale", Computer science department, University of Geneva, Geneva,2006.[29] F. Lotte, M. C<strong>on</strong>gedo, A. Lecuyer, F. Lamarche, <strong>and</strong> B. Arnaldi, "A review ofclassificati<strong>on</strong> algorithms <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG-based brain-computer interfaces," Journal of NeuralEngineering, vol. 4, No. 2, Jun 2007, pp. R1-R13.[30] J. Kr<strong>on</strong>egg, G. Chanel, S. Voloshynovskiy, <strong>and</strong> T. Pun, "EEG-based synchr<strong>on</strong>ized braincomputerinterfaces: A model <str<strong>on</strong>g>for</str<strong>on</strong>g> optimizing the number of mental tasks," IEEETransacti<strong>on</strong>s <strong>on</strong> Neural Systems <strong>and</strong> Rehabilitati<strong>on</strong> Engineering, vol. 15, No. 1, Mar2007, pp. 50-58.[31] A. Choppin, "EEG-<str<strong>on</strong>g>Based</str<strong>on</strong>g> Human Interface <str<strong>on</strong>g>for</str<strong>on</strong>g> Disabled Individuals: <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> Expressi<strong>on</strong>with Neural Networks", Master thesis, In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> processing, Tokyo institute oftechnology, Yokohama, Japan, 2000.[32] R. S. Lazarus, "Progress <strong>on</strong> a Cognitive Motivati<strong>on</strong>al Relati<strong>on</strong>al Theory of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>,"American Psychologist, vol. 46, No. 8, Aug 1991, pp. 819-834.[33] R. R. Cornelius, "Theoretical approaches to emoti<strong>on</strong>," Proc. Int. Speech Communicati<strong>on</strong>Associati<strong>on</strong> (ISCA) Workshop <strong>on</strong> Speech <strong>and</strong> <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>, Belfast, Irel<strong>and</strong>, 2000.[34] A. Ort<strong>on</strong>y <strong>and</strong> T. J. Turner, "Whats Basic About Basic <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s," Psychological Review,vol. 97, No. 3, Jul 1990, pp. 315-331.166


References[35] C. M. Whissell, "The dicti<strong>on</strong>ary of affect in language," in <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>: Theory, Research <strong>and</strong>Experience, vol. 4, R. Pluchik <strong>and</strong> H. Kellerman, Eds. New York: Academic Press, Inc.,1989, pp. 113-131.[36] G. Chanel, J. Kr<strong>on</strong>egg, D. Gr<strong>and</strong>jean, <strong>and</strong> T. Pun, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment: Arousalevaluati<strong>on</strong> using EEG's <strong>and</strong> peripheral physiological signals," in Multimedia C<strong>on</strong>tentRepresentati<strong>on</strong>, Classificati<strong>on</strong> <strong>and</strong> Security, vol. 4105, A. K. J. B. Gunsel, A. M. Tekalp,B. Sankur, Ed., Lecture Notes in Computer Science ed. Istanbul, Turkey: Springer LNCS,2006, pp. 530-537.[37] G. Chanel, C. Rebetez, M. Bétrancourt, <strong>and</strong> T. Pun, "Boredom, Engagement <strong>and</strong> Anxietyas Indicators <str<strong>on</strong>g>for</str<strong>on</strong>g> Adaptati<strong>on</strong> to Difficulty in Games," 12th Internati<strong>on</strong>al MindTrekC<strong>on</strong>ference: Entertainment <strong>and</strong> Media in the Ubiquitous Era, ACM, Tampere, Finl<strong>and</strong>,Oct., 2008.[38] G. Chanel, K. Ansari-Asl, <strong>and</strong> T. Pun, "Valence-arousal evaluati<strong>on</strong> using physiologicalsignals in an emoti<strong>on</strong> recall paradigm," IEEE SMC <strong>and</strong> Internati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong>Systems, Man <strong>and</strong> Cybernetics, Smart cooperative systems <strong>and</strong> cybernetics: advancingknowledge <strong>and</strong> security <str<strong>on</strong>g>for</str<strong>on</strong>g> humanity, M<strong>on</strong>treal, Canada, October 7-10, 2007.[39] G. Chanel, J. J. M. Kierkels, M. Soleymani, <strong>and</strong> T. Pun, "Short-term emoti<strong>on</strong> assessmentin a recall paradigm," accepted <str<strong>on</strong>g>for</str<strong>on</strong>g> publicati<strong>on</strong> in Internati<strong>on</strong>al Journal of Human-Computer Studies, DOI:10.1016/j.ijhcs.2009.03.005, 2009.[40] W. James, "What is an emoti<strong>on</strong> ?," Mind, vol. 9, 1884, pp. 188-205.[41] M. B. Arnold, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> <strong>and</strong> pers<strong>on</strong>ality." New York: [42] P. Ekman, W. V. Friesen, M.O'Sullivan, A. Chan, I. Diacoyanni-Tarlatzis, K. Heider, R. Krause, W. A. LeCompte, T.Pitcairn, <strong>and</strong> P. E. Ricci-Bitti, "Universals <strong>and</strong> cultural differences in the judgments offacial expressi<strong>on</strong>s of emoti<strong>on</strong>," Journal of Pers<strong>on</strong>ality <strong>and</strong> Social Psychology, vol. 53,No. 4, Oct. 1987 1987, pp. 712-717.[43] J. A. Gray, "The Neuropsychology of Anxiety - an Inquiry into the Functi<strong>on</strong>s of theSepto-Hippocampal System," Behavioral <strong>and</strong> <strong>Brain</strong> Sciences, vol. 5, No. 3, 1982, pp.469-484.[44] J. Panksepp, "Toward a General Psycho-Biological Theory of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s," Behavioral <strong>and</strong><strong>Brain</strong> Sciences, vol. 5, No. 3, 1982, pp. 407-422.[45] W. McDougall, An introducti<strong>on</strong> to social psychology. Bost<strong>on</strong>: Luce, 1926.[46] O. H. Mower, Learning theory <strong>and</strong> behavior. New York: Wiley, 1960.[47] R. Pluchik, "A general psychoevoluti<strong>on</strong>ary theory of emoti<strong>on</strong>," in <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>: Theory,Research <strong>and</strong> Experience, vol. 3-31, R. Pluchik <strong>and</strong> H. Kellerman, Eds. New York:Academic Press, Inc., 1980.[48] R. Pluchik, <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>: a psychoevoluti<strong>on</strong>ary synthesis. New York: Harper <strong>and</strong> Row, 1980.[49] R. Pluchik, "A psychoevoluti<strong>on</strong>ary theory of emoti<strong>on</strong>s," Social Science In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>, vol.21, 1982, pp. 529.553.[50] P. Shaver, J. Schwartz, D. Kirs<strong>on</strong>, <strong>and</strong> C. O'C<strong>on</strong>nor, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> Knowledge: FurtherExplorati<strong>on</strong> of a Prototype Approach," in <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s in Social Psychology, W. Parrott, Ed.Philadelphia: Psychology Press, 2001, pp. 26-56.[51] J. A. Russell, "A Circumplex Model of Affect," Journal of Pers<strong>on</strong>ality <strong>and</strong> SocialPsychology, vol. 39, No. 6, Dec. 1980, pp. 1161-1178.[52] J. A. Russel, "Affect Grid: A Single-Item Scale of Pleasure <strong>and</strong> Arousal," Journal ofPers<strong>on</strong>ality <strong>and</strong> Social Psychology, vol. 57, No. 3, Sep. 1989, pp. 493-502.[53] J. A. Russel, M. Lewicka, <strong>and</strong> T. Niit, "A Cross-Cultural Study of a Circumplex Model ofAffect," Journal of Pers<strong>on</strong>ality <strong>and</strong> Social Psychology, vol. 57, No. 5, 1989, pp. 848-856.167


References[54] P. J. Lang, M. M. Bradley, <strong>and</strong> B. N. Cuthbert, "Internati<strong>on</strong>al affective picture system(IAPS): digitized photographs, instructi<strong>on</strong> manual <strong>and</strong> affective ratings," University ofFlorida, Gainesville, FL 2005.[55] M. Isomursu, M. Tahti, S. Vainamo, <strong>and</strong> K. Kuutti, "Experimental evaluati<strong>on</strong> of fivemethods <str<strong>on</strong>g>for</str<strong>on</strong>g> collecting emoti<strong>on</strong>s in field settings with mobile applicati<strong>on</strong>s," Internati<strong>on</strong>alJournal of Human-Computer Studies, vol. 65, No. 4, Apr 2007, pp. 404-418.[56] N. H. Frijda, The emoti<strong>on</strong>s. Cambridge ; New York Cambridge Univ. Press 1986.[57] A. Ort<strong>on</strong>y, G. L. Clore, <strong>and</strong> A. Collins, The cognitive structure of emoti<strong>on</strong>s. Cambridge,MA: Cambridge University Press, 1988.[58] D. S<strong>and</strong>er <strong>and</strong> O. Koenig, "No inferiority Complex in the Study of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> Complexity:A Cognitive Neuroscience Computati<strong>on</strong>al Architecture of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>," Cognitive ScienceQuaterly, vol. 2, No. 3/4, 2002, pp. 249-272.[59] C. Elliott, "The <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Reas<strong>on</strong>er: A process model of emoti<strong>on</strong>s in a multi-agentsystem", Northwestern university, Evanst<strong>on</strong>, Illinois, 1992.[60] C. Adam, B. Gaudou, A. Herzig, <strong>and</strong> D. L<strong>on</strong>gin, "OCC's emoti<strong>on</strong>s: a <str<strong>on</strong>g>for</str<strong>on</strong>g>malizati<strong>on</strong> in aBDI logic," Internati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> Artificial Intelligence: Methodology, Systems,Applicati<strong>on</strong>s (AIMSA 2006), Varna, Bulgaria, Sept 13-15, 2006.[61] D. Gr<strong>and</strong>jean, D. S<strong>and</strong>er, <strong>and</strong> K. R. Scherer, "C<strong>on</strong>scious emoti<strong>on</strong>al experience emerges asa functi<strong>on</strong> of multilevel, appraisal-driven resp<strong>on</strong>se synchr<strong>on</strong>izati<strong>on</strong>," C<strong>on</strong>sciousness <strong>and</strong>Cogniti<strong>on</strong>, vol. 17, No. 2, June 2008, pp. 484-495.[62] S. Delplanque, D. Gr<strong>and</strong>jean, C. Chrea, G. Coppin, L. Aymard, I. Cayeux, C. Margot, M.I. Velazco, D. S<strong>and</strong>er, <strong>and</strong> K. R. Scherer, "Sequential unfolding of novelty <strong>and</strong>pleasantness appraisals of odors: evidence from facial electromyography <strong>and</strong> aut<strong>on</strong>omicreacti<strong>on</strong>s," <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>, In press.[63] A. Ort<strong>on</strong>y, W. Revelle, <strong>and</strong> R. Zinbarg, "Why <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al Intelligence needs a fluidcomp<strong>on</strong>ent," in The science of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al Intelligence, G. Matthews, M. Zeidner, <strong>and</strong> R.D. Roberts, Eds.: Ox<str<strong>on</strong>g>for</str<strong>on</strong>g>d University Press, 2007.[64] K. R. Scherer, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s as Episodes of Subsystem Synchr<strong>on</strong>izati<strong>on</strong> Driven by N<strong>on</strong>linearAppraisal Processes," in <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>, Development <strong>and</strong> Self-Organizati<strong>on</strong>, Dynamic SystemApproaches to <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al Development, M. D. Lewis <strong>and</strong> I. Granic, Eds.: CambridgeUniversity press, 2000.[65] I. Cohen, N. Sebe, A. Garg, L. S. Chen, <strong>and</strong> T. S. Huang, "Facial expressi<strong>on</strong> recogniti<strong>on</strong>from video sequences: temporal <strong>and</strong> static modeling," Computer Visi<strong>on</strong> <strong>and</strong> ImageUnderst<strong>and</strong>ing, vol. 91, No. 1-2, Jul-Aug 2003, pp. 160-187.[66] Z. Zeng, J. Tu, B. M. Pianfetti, <strong>and</strong> T. S. Huang, "Audio–Visual <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Expressi<strong>on</strong>Recogniti<strong>on</strong> Through Multistream Fused HMM," IEEE Transacti<strong>on</strong>s <strong>on</strong> multimedia, vol.10, No. 4, June 2008, pp. 570-577.[67] Z. H. Zeng, M. Pantic, G. I. Roisman, <strong>and</strong> T. S. Huang, "A Survey of Affect Recogniti<strong>on</strong>Methods: Audio, Visual, <strong>and</strong> Sp<strong>on</strong>taneous Expressi<strong>on</strong>s," IEEE Transacti<strong>on</strong>s <strong>on</strong> PatternAnalysis <strong>and</strong> Machine Intelligence, vol. 31, No. 1, Jan 2009, pp. 39-58.[68] Y. Wang <strong>and</strong> L. Guan, "Recognizing human emoti<strong>on</strong>al state from audiovisual signals,"IEEE Trans. <strong>on</strong> multimedia, vol. 10, No. 5, Aug. 2008, pp. 936-946.[69] L. Devillers, L. Vidrascu, <strong>and</strong> L. Lamel, "Challenges in real-life emoti<strong>on</strong> annotati<strong>on</strong> <strong>and</strong>machine learning based detecti<strong>on</strong>," Neural Networks, vol. 18, 2005, pp. 407-422.[70] P. Rani, N. Sarkar, <strong>and</strong> C. Liu, "Maintaining Optimal Challenge in Computer Gamesthrough Real-Time Physiological Feedback," 11th HCI Internati<strong>on</strong>al, Las Vegas, USA,July 22-27, 2005.168


References[71] Z. Hammal, "Segmentati<strong>on</strong> des Traits du Visage, Analyse et Rec<strong>on</strong>naissanced'Expressi<strong>on</strong>s Faciales par les Modèles de Croyance Transférable", Sciences Cognitive,Joseph Fourier Grenoble, Grenoble, 2006.[72] P. Rani <strong>and</strong> N. Sarkar, "Operator Engagement Detecti<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> Robot Behavior Adaptati<strong>on</strong>,"Int. Journal of Advanced Robotic Systems, vol. 4, No. 1, 2007, pp. 1-12.[73] J. D. Morris, "SAM:The Self-<str<strong>on</strong>g>Assessment</str<strong>on</strong>g> Manikin, An Efficient Cross-CulturalMeasurement of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al Resp<strong>on</strong>se," Journal of Advertising Research, vol. 35, No. 6,November 1995.[74] D. Purves, G. J. Augustine, D. Fitzpatrick, <strong>and</strong> L. C. Katz, Neuroscience (french editi<strong>on</strong>),1997.[75] G. Pfurtscheller <strong>and</strong> F. H. L. da Silva, "Event-related EEG/MEG synchr<strong>on</strong>izati<strong>on</strong> <strong>and</strong>desynchr<strong>on</strong>izati<strong>on</strong>: basic principles," Clinical Neurophysiology, vol. 110, No. 11, Nov1999, pp. 1842-1857.[76] R. Oostenveld <strong>and</strong> P. Praamstra, "The five percent electrode system <str<strong>on</strong>g>for</str<strong>on</strong>g> high-resoluti<strong>on</strong>EEG <strong>and</strong> ERP measurements," Clinical Neurophysiology, vol. 112, No. 4, Apr 2001, pp.713-719.[77] J. E. LeDoux, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> Circuits in the <strong>Brain</strong>," Annual Review of Neuroscience, vol. 23,March 2000, pp. 155-184.[78] R. J. Davids<strong>on</strong>, D. C. Jacks<strong>on</strong>, <strong>and</strong> N. H. Kalin, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>, Plasticity, C<strong>on</strong>text, <strong>and</strong>Regulati<strong>on</strong>: Perspectives From <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Neuroscience," Psychological Bulletin, vol. 126,No. 6, 2000, pp. 890-909.[79] E. T. Rolls, "Précis of The brain <strong>and</strong> emoti<strong>on</strong>," Behavioral <strong>and</strong> <strong>Brain</strong> Sciences, vol. 23,No. 2, April 2000, pp. 177-233.[80] R. Adolphs, D. Tranel, <strong>and</strong> A. R. Damasio, "Dissociable neural systems <str<strong>on</strong>g>for</str<strong>on</strong>g> recognizingemoti<strong>on</strong>s," <strong>Brain</strong> <strong>and</strong> Cogniti<strong>on</strong>, vol. 52, No. 1, Jun 2003, pp. 61-69.[81] R. J. Davids<strong>on</strong>, "<str<strong>on</strong>g>Affective</str<strong>on</strong>g> neuroscience <strong>and</strong> psychophysiology: Toward a synthesis,"Psychophysiology, vol. 40, No. 5, 2003, pp. 655-665.[82] L. I. Aftanas, N. V. Reva, A. A. Varlamov, S. V. Pavlov, <strong>and</strong> V. P. Makhnev, "Analysisof Evoked EEG Synchr<strong>on</strong>izati<strong>on</strong> <strong>and</strong> Desynchr<strong>on</strong>izati<strong>on</strong> in C<strong>on</strong>diti<strong>on</strong>s of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>alActivati<strong>on</strong> in Humans: Temporal <strong>and</strong> Topographic Characteristics," Neuroscience <strong>and</strong>Behavioral Physiology, vol. 34, No. 8, 29 May 2002 2004, pp. 859-867.[83] T. Costa, E. Rogn<strong>on</strong>i, <strong>and</strong> D. Galati, "EEG phase synchr<strong>on</strong>izati<strong>on</strong> during emoti<strong>on</strong>alresp<strong>on</strong>se to positive <strong>and</strong> negative film stimuli," Neuroscience Letters, vol. 406, No. 3, Oct9 2006, pp. 159-164.[84] A. R. Damasio, T. J. Grabowski, A. Bechara, H. Damasio, L. L. B. P<strong>on</strong>to, J. Parvizi, <strong>and</strong>R. D. Hichwa, "Subcortical <strong>and</strong> cortical brain activity during the feeling of self-generatedemoti<strong>on</strong>s," Nature Neuroscience, vol. 3, No. 10, 2000, pp. 1049-1056.[85] C. Féré, "Note sur les modificati<strong>on</strong>s de la resistance electrique sous l'influence desexcitati<strong>on</strong>s sensorielles et des emoti<strong>on</strong>s," Compt. Rend. Soc. Biol., vol. 5, 1888, pp. 217-219.[86] D. C. Fowles, "The eccrine system <strong>and</strong> electrodermal activity," in Psychophysiology, M.G. H. Coles, E. D<strong>on</strong>chin, <strong>and</strong> S. W. Porges, Eds. New York: Guil<str<strong>on</strong>g>for</str<strong>on</strong>g>d Press, 1986, pp. 51-96.[87] J. A. Healey, "Wearable <strong>and</strong> Automotive Systems <str<strong>on</strong>g>for</str<strong>on</strong>g> Affect Recogniti<strong>on</strong> fromPhysiology", PhD Thesis, EECS Dpt., MIT, USA, 2000., Electrical Engineering <strong>and</strong>Computer Science Dept., MIT, 2000.169


References[88] M. E. Daws<strong>on</strong>, A. M. Schell, <strong>and</strong> D. L. Fili<strong>on</strong>, "The electrodermal resp<strong>on</strong>se system," inPrinciples of psychophysiology: Physical, social <strong>and</strong> inferential elements, J. T. Cacioppo<strong>and</strong> L. G. Tassinary, Eds. Cambridge: Cambridge University Press, 1990, pp. 295-324.[89] H. Sequeira, P. Hot, L. Silvert, <strong>and</strong> S. Delplanque, "Electrical aut<strong>on</strong>omic correlates ofemoti<strong>on</strong>," Internati<strong>on</strong>al Journal of Psychophysiology, No. In press, 2008.[90] S. Delplanque, D. Gr<strong>and</strong>jean, C. Chrea, L. Aymard, I. Cayeux, B. L. Calve, M. I.Velazco, K. R. Scherer, <strong>and</strong> D. S<strong>and</strong>er, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al Processing of Odors: Evidence <str<strong>on</strong>g>for</str<strong>on</strong>g> aN<strong>on</strong>linear Relati<strong>on</strong> between Pleasantness <strong>and</strong> Familiarity Evaluati<strong>on</strong>s," Chemical Senses,vol. 33, 2008, pp. 469-479.[91] D. Shapiro, L. D. Jamner, J. D. Lane, K. C. Light, M. Myrtek, Y. Sawada, <strong>and</strong> A. Steptoe,"Blood pressure publicati<strong>on</strong> guidelines," Psychophysiology, vol. 33, No. 1, Jan 1996, pp.1-12.[92] R. Sinha, W. R. Lovallo, <strong>and</strong> O. A. Pars<strong>on</strong>s, "Cardiovascular differentiati<strong>on</strong> of emoti<strong>on</strong>s,"Psychosomatic Medicine, vol. 54, No. 4, 1992, pp. 422-435.[93] C. L. Lisetti <strong>and</strong> F. Nasoz, "Using N<strong>on</strong>invasive Wearable Computers to RecognizeHuman <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s from Physiological Signals," Journal <strong>on</strong> Applied Signal Processing,vol. 11, 2004, pp. 1672-1687.[94] R. A. Wright, R. J. C<strong>on</strong>trada, <strong>and</strong> M. J. Patane, "Task-Difficulty, Cardiovascular-Resp<strong>on</strong>se, <strong>and</strong> the Magnitude of Goal Valence," Journal of Pers<strong>on</strong>ality <strong>and</strong> SocialPsychology, vol. 51, No. 4, Oct 1986, pp. 837-843.[95] R. A. Wright <strong>and</strong> J. C. Dill, "Blood-Pressure Resp<strong>on</strong>ses <strong>and</strong> Incentive Appraisals as aFuncti<strong>on</strong> of Perceived Ability <strong>and</strong> Objective Task Dem<strong>and</strong>," Psychophysiology, vol. 30,No. 2, Mar 1993, pp. 152-160.[96] G. G. Bernts<strong>on</strong>, J. T. Bigger, D. L. Eckberg, P. Grossman, P. G. Kaufmann, M. Malik, H.N. Nagaraja, S. W. Porges, J. P. Saul, P. H. St<strong>on</strong>e, <strong>and</strong> M. W. V<strong>and</strong>erMolen, "Heart ratevariability: origins, methods, <strong>and</strong> interpretive caveats," Psychophysiology, vol. 34, No. 6,Nov 1997, pp. 623-648.[97] J. R. Jennings, W. K. Berg, J. S. Hutches<strong>on</strong>, P. Obrist, S. Porges, <strong>and</strong> G. Turpin,"Publicati<strong>on</strong> Guidelines <str<strong>on</strong>g>for</str<strong>on</strong>g> Heart-Rate Studies in Man," Psychophysiology, vol. 18, No.3, 1981, pp. 226-231.[98] T. Ritz, B. Dahme, A. B. Dubois, H. Folgering, G. K. Fritz, A. Harver, H. Kotses, P. M.Lehrer, C. Ring, A. Steptoe, <strong>and</strong> K. P. van de Woestijne, "Guidelines <str<strong>on</strong>g>for</str<strong>on</strong>g> mechanical lungfuncti<strong>on</strong> measurements in psychophysiology," Psychophysiology, vol. 39, No. 5, Sep2002, pp. 546-567.[99] F. H. Wilhelm, M. C. Pfaltz, <strong>and</strong> P. Grossman, "C<strong>on</strong>tinuous electr<strong>on</strong>ic data capture ofphysiology, behavior <strong>and</strong> experience in real life: towards ecological momentaryassessment of emoti<strong>on</strong>," Interacting with Computers, vol. 18, No. 2, 2006, pp. 171-186.[100] J. Kim, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> Recogniti<strong>on</strong> from Physiological Measurement," Humaine EuropeanNetwork of Excellence Workshop, Santorini, Greece, September 18-21, 2004.[101] A. C. Guyt<strong>on</strong> <strong>and</strong> J. E. Hall, Textbook of medical physiology 11th editi<strong>on</strong>: Elsevier Inc.,2006.[102] R. A. McFarl<strong>and</strong>, "Relati<strong>on</strong>ship of skin temperature changes to the emoti<strong>on</strong>saccompanying music," Applied Psychophysiology <strong>and</strong> Biofeedback, vol. 10, No. 3, sept.1985, pp. 255-267.[103] C. Puri, L. Ols<strong>on</strong>, I. Pavlidis, J. Levine, <strong>and</strong> J. Starren, "StressCam: n<strong>on</strong>-c<strong>on</strong>tactmeasurement of users’ emoti<strong>on</strong>al states through thermal imaging," C<strong>on</strong>ference <strong>on</strong> HumanFactors in <str<strong>on</strong>g>Computing</str<strong>on</strong>g> Systems (CHI’05), Portl<strong>and</strong>, Oreg<strong>on</strong>, April 2-7, 2005.170


References[104] I. Pavlidis, J. Levine, <strong>and</strong> P. Baukol, "Thermal image analysis <str<strong>on</strong>g>for</str<strong>on</strong>g> anxiety detecti<strong>on</strong>,"IEEE Internati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> Image Processing, Thessal<strong>on</strong>iki, Greece, Oct. 7-10,2001.[105] S. Jarlier, D. Gr<strong>and</strong>jean, K. N’Diaye, S. Delplanque, D. S<strong>and</strong>er, P. Vuilleumier, <strong>and</strong> K. R.Scherer, "Thermal imaging of facial expressi<strong>on</strong>s: investigating thermal correlates ofFacial Acti<strong>on</strong> Units activities," 10th Internati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> Cognitive Neuroscience,Bodrum, Turkey, Sept. 1-5, 2008.[106] R. W. Picard, E. Vyzas, <strong>and</strong> J. Healey, "Toward machine emoti<strong>on</strong>al intelligence: Analysisof affective physiological state," IEEE Transacti<strong>on</strong>s <strong>on</strong> Pattern Analysis <strong>and</strong> MachineIntelligence, vol. 23, No. 10, Oct 2001, pp. 1175-1191.[107] A. Haag, S. Gor<strong>on</strong>zy, P. Schaich, <strong>and</strong> J. Williams, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> Recogniti<strong>on</strong> Using Bio-Sensors: First Step Toward an Automatic System," <str<strong>on</strong>g>Affective</str<strong>on</strong>g> Dialog Systems: Tutorial<strong>and</strong> Research Workshop, Kloster Irsee, Germany, June 14-16, 2004.[108] G. Stemmler, M. Heldmann, C. A. Pauls, <strong>and</strong> T. Scherer, "C<strong>on</strong>straints <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>specificity in fear <strong>and</strong> anger: The c<strong>on</strong>text counts," Psychophysiology, vol. 38, No. 2, Mar2001, pp. 275-291.[109] C. D. Katsis, N. Katertsidis, G. Ganiatras, <strong>and</strong> D. I. Fotiadis, "Toward emoti<strong>on</strong>recogniti<strong>on</strong> in car racing drivers: a biosignal processing approach," IEEE Trans. Systems,Man <strong>and</strong> Cybernetics - Part A: Systems <strong>and</strong> Humans, vol. 38, No. 3, May 2008, pp. 502-512.[110] K. H. Kim, S. W. Bang, <strong>and</strong> S. R. Kim, "<str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> recogniti<strong>on</strong> system using short termm<strong>on</strong>itoring of physiological signals," Medical Biological Engineering <strong>and</strong> computing,vol. 42, 2004, pp. 419-427.[111] J. Wagner, J. Kim, <strong>and</strong> E. André, "From physiological signals to emoti<strong>on</strong>s: implementing<strong>and</strong> comparing selected methods <str<strong>on</strong>g>for</str<strong>on</strong>g> features extracti<strong>on</strong> <strong>and</strong> classificati<strong>on</strong>," IEEEInternati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> Multimedia & Expo, 6-8 July, 2005.[112] R. Sinha <strong>and</strong> O. A. Pars<strong>on</strong>s, "Multivariate resp<strong>on</strong>se patterning of fear <strong>and</strong> anger,"Cogniti<strong>on</strong> & <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>, vol. 10, No. 2, Mar 1996, pp. 173-198.[113] K. Takahashi, "Remarks <strong>on</strong> <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> Recogniti<strong>on</strong> from Bio-Potential Signals,"Proceedings of the 2nd Internati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> Aut<strong>on</strong>omous Robots <strong>and</strong> Agents,Palmerst<strong>on</strong> North, New Zeal<strong>and</strong>, December 13-15, 2004.[114] E. Le<strong>on</strong>, G. Clarke, V. Callaghan, <strong>and</strong> F. Sepulveda, "A user-independent real-timeemoti<strong>on</strong> recogniti<strong>on</strong> system <str<strong>on</strong>g>for</str<strong>on</strong>g> software agents in domestic envir<strong>on</strong>ments," EngineeringApplicati<strong>on</strong>s of Artificial Intelligence, vol. 20, No. 3, Apr 2007, pp. 337-345.[115] T. Sakata, S. Watanuki, H. Sakamoto, T. Sumi, <strong>and</strong> Y.-K. Kim, "Objective evaluati<strong>on</strong> ofKansei by a complementary use of physiological indexes, brain wave <strong>and</strong> facialexpressi<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> user oriented designs," Proceedings of the10th Qmod c<strong>on</strong>ference, QualityManagement <strong>and</strong> Organisati<strong>on</strong>al Development: Our Dreams of Excellence, Helsingborg,Sweden, June 18-20, 2007.[116] P. Rani, C. Liu, <strong>and</strong> N. Sarkar, "An Empirical study of machine learning techniques <str<strong>on</strong>g>for</str<strong>on</strong>g>affect recogniti<strong>on</strong> in human-robot interacti<strong>on</strong>," Pattern Analysis & Applicati<strong>on</strong>s, vol. 9,2006, pp. 58-69.[117] M. Pelzer, J. D. Schipke, D. Horstkotte, <strong>and</strong> G. Arnold, "Effect of Respirati<strong>on</strong> <strong>on</strong> Short-Term Heart-Rate-Variability," Faseb Journal, vol. 8, No. 5, Mar 18 1994, pp. A846.[118] R. Bailón, P. Laguna, L. Mainardi, <strong>and</strong> L. Sörnmo, "Analysis of Heart Rate VariabilityUsing Time-Varying Frequency B<strong>and</strong>s <str<strong>on</strong>g>Based</str<strong>on</strong>g> <strong>on</strong> Respiratory Frequency," IEEE 29thInternati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> EMBS, Ly<strong>on</strong>, France, Aug. 23-26, 2007.171


References[119] R. O. Duda, P. E. Hart, <strong>and</strong> D. G. Stork, Pattern Classificati<strong>on</strong>, sec<strong>on</strong>d ed: WileyInterscience, 2001.[120] C. M. Bishop, Pattern recogniti<strong>on</strong> <strong>and</strong> machine learning: Springer, 2006.[121] E. L. van den Broek, M. H. Schut, J. H. D. M. Westerink, J. van Herk, <strong>and</strong> K.Tuinenbreijer, "<str<strong>on</strong>g>Computing</str<strong>on</strong>g> emoti<strong>on</strong> awareness through facial electromyography,"Computer Visi<strong>on</strong> in Human-Computer Interacti<strong>on</strong>, vol. 3979, 2006, pp. 52-63.[122] K. Takahashi, "Remarks <strong>on</strong> SVM-based emoti<strong>on</strong> recogniti<strong>on</strong> from multi-modal biopotentialsignals," Proceedings of the 13th IEEE Internati<strong>on</strong>al Workshop <strong>on</strong> Robot <strong>and</strong>Human Interactive Communicati<strong>on</strong> Sept. 20-22, 2004.[123] B. Herbelin, P. Benzaki, F. Riquier, O. Renault, <strong>and</strong> D. Thalmann, "Using physiologicalmeasures <str<strong>on</strong>g>for</str<strong>on</strong>g> emoti<strong>on</strong>al assessment: a computer-aided tool <str<strong>on</strong>g>for</str<strong>on</strong>g> cognitive <strong>and</strong> behaviouraltherapy," 5th Internati<strong>on</strong>al C<strong>on</strong>ference <strong>on</strong> Disability, Ox<str<strong>on</strong>g>for</str<strong>on</strong>g>d, 2004.[124] C. Lee, S. Yoo, Y. Park, N. Kim, K. Je<strong>on</strong>g, <strong>and</strong> B. Lee, "Using Neural Network toRecognize Human <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s from Heart Rate Variability <strong>and</strong> Skin Resistance,"Engineering in Medecine <strong>and</strong> Biology 27th Annual C<strong>on</strong>ference, Shangai, China,September 1-4, 2005.[125] R. L. M<strong>and</strong>ryk, K. M. Inkpen, <strong>and</strong> T. W. Calvert, "Using psychophysiological techniquesto measure user experience with entertainment technologies," Behaviour & In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>Technology, vol. 25, No. 2, Mar-Apr 2006, pp. 141-158.[126] Biosemi, "http://www.biosemi.com/. Biosemi website 30/01/2009."[127] D. C. Fowles, M. J. Christie, R. Edelberg, W. W. Grings, D. T. Lykken, <strong>and</strong> P. H.Venables, "Publicati<strong>on</strong> Recommendati<strong>on</strong>s <str<strong>on</strong>g>for</str<strong>on</strong>g> Electrodermal Measurements,"Psychophysiology, vol. 18, No. 3, May 1981, pp. 232-239.[128] M. Aboy, J. McNames, T. Th<strong>on</strong>g, D. Tsunami, M. S. Ellenby, <strong>and</strong> B. Goldstein, "Anautomatic beat detecti<strong>on</strong> algorithm <str<strong>on</strong>g>for</str<strong>on</strong>g> pressure signals," IEEE Transacti<strong>on</strong>s <strong>on</strong>Biomedical Engineering, vol. 52, No. 10, Oct 2005, pp. 1662-1670.[129] K. Ansari-Asl, G. Chanel, <strong>and</strong> T. Pun, "A channel selecti<strong>on</strong> method <str<strong>on</strong>g>for</str<strong>on</strong>g> EEGclassificati<strong>on</strong> in emoti<strong>on</strong> assessment based <strong>on</strong> synchr<strong>on</strong>izati<strong>on</strong> likelihood," 15th EuropeanSignal Processing C<strong>on</strong>ference (Eusipco 2007), Poznan, Pol<strong>and</strong>, Sept. 3-7, 2007.[130] C. Berka, D. J. Levendowski, M. M. Cvetinovic, M. M. Petrovic, G. Davis, M. N.Lumicao, V. T. Zivkovic, M. V. Popovic, <strong>and</strong> R. Olmstead, "Real-time analysis of EEGindexes of alertness, cogniti<strong>on</strong>, <strong>and</strong> memory acquired with a wireless EEG headset,"Internati<strong>on</strong>al Journal of Human-Computer Interacti<strong>on</strong>, vol. 17, No. 2, 2004, pp. 151-170.[131] A. T. Pope, E. H. Bogart, <strong>and</strong> D. S. Bartolome, "Biocybernetic System Evaluates Indexesof Operator Engagement in Automated Task," Biological Psychology, vol. 40, No. 1-2,May 1995, pp. 187-195.[132] M. Besserve, M. Philippe, G. Florence, F. Laurent, L. Garnero, <strong>and</strong> J. Martinerie,"Predicti<strong>on</strong> of per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance level during a cognitive task from <strong>on</strong>going EEG oscillatoryactivites," Clinical Neurophysiology, vol. 119, No. 4, April 2008, pp. 897-908.[133] F. G. Freeman, P. J. Mikulka, M. W. Scerbo, <strong>and</strong> L. Scott, "An evaluati<strong>on</strong> of an adaptiveautomati<strong>on</strong> system using a cognitive vigilance task," Biological Psychology, vol. 67, No.3, Nov. 2004, pp. 283-297.[134] R. Moddemeijer, "On Estimati<strong>on</strong> of Entropy <strong>and</strong> Mutual In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> of C<strong>on</strong>tinuousDistributi<strong>on</strong>s," Signal Processing, vol. 16, No. 3, Mar 1989, pp. 233-248.[135] E. Pauwels, "Ethics <str<strong>on</strong>g>for</str<strong>on</strong>g> researchers, facilitating research excellence in FP7," 2007.[136] R. Kohavi, "A Study of Cross-Validati<strong>on</strong> <strong>and</strong> Bootstrap <str<strong>on</strong>g>for</str<strong>on</strong>g> Accuracy Estimati<strong>on</strong> <strong>and</strong>Model Selecti<strong>on</strong>," 14th IJCAI, M<strong>on</strong>tréal, Canada, August 20-25, 1995.172


References[137] C. Cortes <strong>and</strong> V. Vapnik, "Support-Vector Networks," Machine Learning, vol. 20, No. 3,Sep 1995, pp. 273-297.[138] J. Hua, Z. Xi<strong>on</strong>g, J. Lowey, E. Suh, <strong>and</strong> E. R. Dougherty, "Optimal number of features asa functi<strong>on</strong> of sample size <str<strong>on</strong>g>for</str<strong>on</strong>g> various classificati<strong>on</strong> rules," Bioin<str<strong>on</strong>g>for</str<strong>on</strong>g>matics, vol. 21, No. 8,2005, pp. 1509-1515.[139] J. Platt, "Probabilities <str<strong>on</strong>g>for</str<strong>on</strong>g> SV Machines," in Advances in Large Margin Classifiers, A.Smola, P. Bartlett, B. Scholkopf, <strong>and</strong> D. Schuurmans, Eds. Cambridge, MA: MIT Press,2000, pp. 61-64.[140] H. T. Lin, C. J. Lin, <strong>and</strong> R. C. Weng, "A note <strong>on</strong> Platt's probabilistic outputs <str<strong>on</strong>g>for</str<strong>on</strong>g> supportvector machines," Machine Learning, vol. 68, No. 3, Oct 2007, pp. 267-276.[141] T. F. Wu, C. J. Lin, <strong>and</strong> R. C. Weng, "Probability estimates <str<strong>on</strong>g>for</str<strong>on</strong>g> multi-class classificati<strong>on</strong>by pairwise coupling," Journal of Machine Learning Research, vol. 5, Aug 2004, pp. 975-1005.[142] C.-C. Chang <strong>and</strong> C.-J. Lin, "LIBSVM : a library <str<strong>on</strong>g>for</str<strong>on</strong>g> support vector machines," 2001.[143] M. E. Tipping, "Sparse Bayesian learning <strong>and</strong> the relevance vector machine," Journal ofMachine Learning Research, vol. 1, No. 3, Sum 2001, pp. 211-244.[144] H. Zhang <strong>and</strong> J. Malik, "Selecting Shape Features Using Multi-class Relevance VectorMachine," EECS Department, University of Cali<str<strong>on</strong>g>for</str<strong>on</strong>g>nia, Berkeley UCB/EECS-2005-6,October 10 2005.[145] J. Ye, T. Li, T. Xi<strong>on</strong>g, <strong>and</strong> R. Janardan, "Using Uncorrelated Discriminant Analysis <str<strong>on</strong>g>for</str<strong>on</strong>g>Tissue Classificati<strong>on</strong> with Gene Expressi<strong>on</strong> Data," IEEE/ACM Transacti<strong>on</strong>s <strong>on</strong>computati<strong>on</strong>al biology <strong>and</strong> bioin<str<strong>on</strong>g>for</str<strong>on</strong>g>matics, vol. 1, No. 4, Oct-Dec 2004, pp. 181-190.[146] J. Yang, J. Y. Yang, <strong>and</strong> D. Zhang, "What's wr<strong>on</strong>g with Fisher criteri<strong>on</strong>?," PatternRecogniti<strong>on</strong>, vol. 35, No. 11, Nov 2002, pp. 2665-2668.[147] G. H. John, R. Kohavi, <strong>and</strong> K. Pfleger, "Irrelevant Features <strong>and</strong> the Subset Selecti<strong>on</strong>Problem," Proc. 11th Int. C<strong>on</strong>f. <strong>on</strong> Machine Learning, San Francisco, USA, 1994.[148] P. Pudil, F. Ferri, J. Novovicová, <strong>and</strong> J. Kittler, "Floating search methods <str<strong>on</strong>g>for</str<strong>on</strong>g> featureselecti<strong>on</strong> with n<strong>on</strong>m<strong>on</strong>ot<strong>on</strong>ic criteri<strong>on</strong> functi<strong>on</strong>s," Proc. of the IEEE Intl. C<strong>on</strong>f. <strong>on</strong> PatternRecogniti<strong>on</strong>, vol. 2, 1994, pp. 279-283.[149] L. Yu <strong>and</strong> H. Liu, "Efficient Feature Selecti<strong>on</strong> via Analysis of Relevance <strong>and</strong>Redundancy," Journal of Machine Learning Research, vol. 5, 1205-1224 2004.[150] P. Viola <strong>and</strong> M. J<strong>on</strong>es, "Fast <strong>and</strong> robust classificati<strong>on</strong> using asymmetric AdaBoost <strong>and</strong> adetector cascade," Advances in Neural In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> Processing Systems 14, Vols 1 <strong>and</strong> 2,vol. 14, 2002, pp. 1311-1318.[151] J. Kludas, E. Bruno, <strong>and</strong> S. March<strong>and</strong>-Maillet, "Can Feature In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> Interacti<strong>on</strong> help<str<strong>on</strong>g>for</str<strong>on</strong>g> In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> Fusi<strong>on</strong> in Multimedia Problems?," Multimedia Tools <strong>and</strong> Applicati<strong>on</strong>sJournal special issue <strong>on</strong> "Metadata Mining <str<strong>on</strong>g>for</str<strong>on</strong>g> Image Underst<strong>and</strong>ing" vol. 42, No. 1,2009, pp. 57-71.[152] A. A. Freitas, "Underst<strong>and</strong>ing the crucial role of attribute interacti<strong>on</strong> in data mining,"Artificial Intelligence Review, vol. 16, No. 3, Nov 2001, pp. 177-199.[153] C. S<strong>and</strong>ers<strong>on</strong> <strong>and</strong> K. K. Paliwal, "Identity verificati<strong>on</strong> using speech <strong>and</strong> facein<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong>," Digital Signal Processing, vol. 14, No. 5, Sep 2004, pp. 449-480.[154] D. Ruta <strong>and</strong> B. Gabrys, "An Overview of Classifier Fusi<strong>on</strong> Methods," <str<strong>on</strong>g>Computing</str<strong>on</strong>g> <strong>and</strong>In<str<strong>on</strong>g>for</str<strong>on</strong>g>mati<strong>on</strong> Systems, vol. 7, No. 1, Feb. 2000.[155] G. Bologna, B. Deville, <strong>and</strong> T. Pun, "On the use of the auditory pathway to representimage scenes in real-time," Neurocomputing, vol. 72, No. 4-6, Jan 2009, pp. 839-849.173


References[156] K. R. Scherer, E. S. Dan, <strong>and</strong> A. Flykt, "What Determines a Feeling’s Positi<strong>on</strong> in<str<strong>on</strong>g>Affective</str<strong>on</strong>g> Space? A Case <str<strong>on</strong>g>for</str<strong>on</strong>g> Appraisal," Cogniti<strong>on</strong> <strong>and</strong> emoti<strong>on</strong>, vol. 20, No. 1, 2006, pp.92-113.[157] L. Collet <strong>and</strong> R. Duclaux, "Hemispheric Lateralizati<strong>on</strong> of <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>s - Absence ofElectrophysiological Arguments," Physiology & Behavior, vol. 40, No. 2, 1987, pp. 215-220.[158] A. P. R. Smith, K. E. Stephhan, M. D. Rugg, <strong>and</strong> R. J. Dolan, "Task <strong>and</strong> C<strong>on</strong>tentModulate Amygdala-Hippocampal C<strong>on</strong>nectivity in <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g>al Retrieval," Neur<strong>on</strong>, vol. 49,No. 4, Feb. 2006 2006, pp. 631-638.[159] G. Barret, "Event-related potentials (ERPs) as a measure of complex cognitive functi<strong>on</strong>,"Electroencephalography <strong>and</strong> clinical neurophysiology, Supplement, vol. 46, 1996, pp.53–63.[160] C. Guger, G. Edlinger, W. Harkam, I. Niedermayer, <strong>and</strong> G. Pfurtscheller, "How manypeople are able to operate an EEG-based brain-computer interface (BCI)?," IEEETransacti<strong>on</strong>s <strong>on</strong> Neural Systems <strong>and</strong> Rehabilitati<strong>on</strong> Engineering, vol. 11, No. 2, Jun2003, pp. 145-147.[161] M. Csikszentmihalyi, Flow: the psychology of optimal experience: New York: HarperCollins, 1991.[162] K. Salen <strong>and</strong> E. Zimmerman, Rules of play: game design fundamentals. Cambridge: MITPress, 2004.[163] G. H. E. Gendolla <strong>and</strong> M. Richter, "Ego involvement <strong>and</strong> ef<str<strong>on</strong>g>for</str<strong>on</strong>g>t: Cardiovascular,electrodermal, <strong>and</strong> per<str<strong>on</strong>g>for</str<strong>on</strong>g>mance effects," Psychophysiology, vol. 42, No. 5, 2005, pp. 596-603.[164] T. I. Alecu, "Robust Focalized <strong>Brain</strong> Activity Rec<strong>on</strong>structi<strong>on</strong> usingElectroEncephaloGrams", Computer science department, University of Geneva, Geneva,2005.174


List of figuresList of figuresFigure 1.1. Including emoti<strong>on</strong>s in the human-machine loop .............................................................. 2Figure 1.2. <str<strong>on</strong>g>Emoti<strong>on</strong></str<strong>on</strong>g> assessment in human computer interfaces, adapted from the executi<strong>on</strong> /evaluati<strong>on</strong> model [8]. ............................................................................................................................ 4Figure 2.1. Plutchik’s wheel of emoti<strong>on</strong>s. (Left) The basic emoti<strong>on</strong> represented as quadrants <strong>and</strong>possible combinati<strong>on</strong>s of basic emoti<strong>on</strong>s. (Right) The same wheel with the added c<strong>on</strong>cept ofintensity 2 . ............................................................................................................................................. 16Figure 2.2. Valence arousal space with associated labels as (a) points (adapted from [Russell],(adjectives have been changed to nouns <strong>and</strong> <strong>on</strong>ly some of the words are displayed <str<strong>on</strong>g>for</str<strong>on</strong>g> clarity) <strong>and</strong>(b) areas. .............................................................................................................................................. 17Figure 2.3. Self-assessments distributi<strong>on</strong> obtained when eliciting emoti<strong>on</strong>s with images: most ofthe self-assessed images lie inside the U-shape. ............................................................................... 18Figure 2.4. The OCC typology (from [52]). Green <strong>and</strong> blue dotted lines corresp<strong>on</strong>d to theexamples above. .................................................................................................................................. 20Figure 2.5. Picture of the SAM scales (from [68]).The first line evaluates valence from positive(left) to negative (right), the sec<strong>on</strong>d arousal from excited to calm <strong>and</strong> the third dominance fromsubmissive to powerful. ...................................................................................................................... 26Figure 2.6. An acquisiti<strong>on</strong> system <str<strong>on</strong>g>for</str<strong>on</strong>g> visualizati<strong>on</strong> <strong>and</strong> storage of physiological data. ................ 27Figure 2.7. (a) Figure of a neur<strong>on</strong> c<strong>on</strong>nected with two input neur<strong>on</strong>s (named 1 <strong>and</strong> 2). (b)Representati<strong>on</strong> of the integrati<strong>on</strong> of input acti<strong>on</strong> potentials; the neur<strong>on</strong> fires <strong>on</strong>ly if its membranepotential exceeds a given threshold. .................................................................................................. 28Figure 2.8. Image of the brain, the brain stem <strong>and</strong> the cerebellum with the different lobeshighlighted (from [69]). ...................................................................................................................... 29Figure 2.9. Principal structures of the limbic system together with their functi<strong>on</strong>s. ...................... 32Figure 2.10. (left) Example of a signal representing the changes of resistance of the skin, (right)the characterizati<strong>on</strong> of an electrodermal resp<strong>on</strong>se. ........................................................................... 35Figure 2.11. Examples of signals obtained from a respirati<strong>on</strong> belt tied across the chest <strong>and</strong> atemperature sensor placed bellow the nostrils during different type of respirati<strong>on</strong>s. ..................... 39Figure 3.1. Hardware <strong>and</strong> software <str<strong>on</strong>g>for</str<strong>on</strong>g> signal acquisiti<strong>on</strong>. ............................................................... 51175


List of figuresFigure 3.2. (left) A participant wearing the EEG cap with 64 electrodes plugged. (right) Top headview with the positi<strong>on</strong>s <strong>and</strong> names of the 64 electrodes used <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG recording. For a 19electrodes c<strong>on</strong>figurati<strong>on</strong> <strong>on</strong>ly the green electrodes were used. ........................................................ 53Figure 3.3. Pictures <strong>and</strong> positi<strong>on</strong>s of the sensors used to m<strong>on</strong>itor peripheral activity. The CMS /DRL positi<strong>on</strong> was used <strong>on</strong>ly in the case where EEG activity was not m<strong>on</strong>itored simultaneouslywith peripheral activity. ...................................................................................................................... 54Figure 3.4. The heart waves in a BVP signal. (Left) Three pulses of the BVP signal with thedifferent peaks, (rigth) example of a pulse where it is difficult to identify the different peaks. .... 57Figure 3.5. Example of the beat detecti<strong>on</strong> <strong>and</strong> HR computati<strong>on</strong> algorithm <strong>on</strong> a 9 sec<strong>on</strong>ds signal.The HR signal is represented as a staircase functi<strong>on</strong> with the length of a step corresp<strong>on</strong>ding to thedurati<strong>on</strong> of an IBI. ............................................................................................................................... 58Figure 3.6. Top head view with EEG electrode locati<strong>on</strong>s <strong>and</strong> corresp<strong>on</strong>ding frequency b<strong>and</strong>s(from [77]). .......................................................................................................................................... 63Figure 4.1. Validati<strong>on</strong> scheme <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong>, where ŷ is the vector of the classes estimated bymodel <str<strong>on</strong>g>for</str<strong>on</strong>g> the test set, A is the accuracy............................................................................................. 74Figure 4.2. Obtaining posterior probabilities p( i | h) from SVM outputs. a) Histogramsrepresenting the distributi<strong>on</strong>s of the SVM output <str<strong>on</strong>g>for</str<strong>on</strong>g> two classes. b) Posterior probabilitiesestimates from the Bayes rules applied <strong>on</strong> the histogram of a) <strong>and</strong> from the sigmoid fit proposedby Platt [135]. ...................................................................................................................................... 79Figure 4.3. Different possible distributi<strong>on</strong>s of a feature value <str<strong>on</strong>g>for</str<strong>on</strong>g> a 3 classes scenario (green, red<strong>and</strong> black classes). (left) The feature is relevant since it is usefull to distinguish the green classfrom the others (low p value). (right) A n<strong>on</strong> relevant feature (high p value). ................................. 81Figure 5.1. Descripti<strong>on</strong> of the acquisiti<strong>on</strong> protocol. (left) the modified SAM used <str<strong>on</strong>g>for</str<strong>on</strong>g> selfassessment. (right) the schedule of the protocol. .............................................................................. 91Figure 5.2. Histograms of the IAPS <strong>and</strong> self evaluati<strong>on</strong>s (valence <strong>and</strong> arousal) <str<strong>on</strong>g>for</str<strong>on</strong>g> the valenceexperiment. For easier comparis<strong>on</strong> of IAPS evaluati<strong>on</strong>s <strong>and</strong> self evaluati<strong>on</strong>s the IAPS valueshave been normalized to the same range as the self evaluati<strong>on</strong>s. .................................................... 94Figure 5.3. Histograms of the IAPS <strong>and</strong> self evaluati<strong>on</strong>s (valence <strong>and</strong> arousal) <str<strong>on</strong>g>for</str<strong>on</strong>g> the arousalexperiment. For easier comparis<strong>on</strong> of IAPS evaluati<strong>on</strong>s <strong>and</strong> self evaluati<strong>on</strong>s the IAPS valueshave been normalized in the same range as the self evaluati<strong>on</strong>s. .................................................... 95Figure 5.4. LDA accuracy <str<strong>on</strong>g>for</str<strong>on</strong>g> classificati<strong>on</strong> of negative <strong>and</strong> positive stimuli. ............................... 98176


List of figuresFigure 5.5. Classifiers accuracy with 2 classes c<strong>on</strong>structed from self-assessment. ........................ 99Figure 5.6. Classifiers accuracy with 3 classes c<strong>on</strong>structed from self-assessment. ...................... 100Figure 6.1. (left) The different emoti<strong>on</strong>al classes represented in the valence-arousal space <strong>and</strong>their associated image. (right) schedule of the protocol <strong>and</strong> detail of a trial. ............................... 103Figure 6.2. Complete process of trial acquisiti<strong>on</strong>, classificati<strong>on</strong>, fusi<strong>on</strong> <strong>and</strong> rejecti<strong>on</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> a givenparticipant. As defined in Chapter 4, k i is the c<strong>on</strong>fidence measure of class i after opini<strong>on</strong> fusi<strong>on</strong><strong>and</strong> reject is the rejecti<strong>on</strong> threshold. .................................................................................................. 105Figure 6.3. Mean classifier accuracy across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG_STFT feature set <strong>and</strong> thedifferent classificati<strong>on</strong> schemes. The bars <strong>on</strong> top of each column represents the st<strong>and</strong>ard deviati<strong>on</strong>across participants. ............................................................................................................................ 110Figure 6.4. Mean classifier accuracy across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> the EEG_MI feature set <strong>and</strong> thedifferent classificati<strong>on</strong> schemes. The bars <strong>on</strong> top of each column represents the st<strong>and</strong>ard deviati<strong>on</strong>across participants. ............................................................................................................................ 111Figure 6.5. Mean classifier accuracy across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features <strong>and</strong> the differentclassificati<strong>on</strong> schemes. The bars <strong>on</strong> top of each column represents the st<strong>and</strong>ard deviati<strong>on</strong> acrossparticipants. ....................................................................................................................................... 111Figure 6.6. classificati<strong>on</strong> accuracy using participant 1 EEG_STFT features with LDA <strong>and</strong> withSVM <strong>on</strong> the five sets of classes, with or without FCBF feature selecti<strong>on</strong>. The bottom horiz<strong>on</strong>talaxis indicates the value of the threshold FCBF , while the top horiz<strong>on</strong>tal axis corresp<strong>on</strong>ds to thenumber of selected features. ............................................................................................................. 115Figure 6.7. Accuracy of LDA <strong>and</strong> the Linear SVM classifiers <str<strong>on</strong>g>for</str<strong>on</strong>g> different numbers of selectedfeatures of the EEG_STFT feature set using the Fisher criteri<strong>on</strong> (<strong>on</strong>ly <str<strong>on</strong>g>for</str<strong>on</strong>g> the CPN classificati<strong>on</strong>scheme). Only the number of features marked with a ‘+’ have been computed while the othervalues are linearly interpolated. ....................................................................................................... 116Figure 6.8. Average accuracy across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> different modalities <strong>and</strong> their associatedclassifiers, as well as <str<strong>on</strong>g>for</str<strong>on</strong>g> fusi<strong>on</strong> of the two EEG <strong>and</strong> the three physiological modalities. ........... 117Figure 6.9. Relati<strong>on</strong> between the threshold value, classificati<strong>on</strong> accuracy <strong>and</strong> the amount ofeliminated samples <str<strong>on</strong>g>for</str<strong>on</strong>g> the CPN classificati<strong>on</strong> task. ....................................................................... 118Figure 7.1. Flow chart <strong>and</strong> the suggested automatic adaptati<strong>on</strong> to emoti<strong>on</strong>al reacti<strong>on</strong>s. ............. 122Figure 7.2. Screen shot of the Tetris (DotNETris) game................................................................ 123177


List of figuresFigure 7.3. Histogram of the skill levels of the 20 participants. .................................................... 124Figure 7.4. Schedule of the protocol. ............................................................................................... 125Figure 7.5. Mean <strong>and</strong> st<strong>and</strong>ard deviati<strong>on</strong> of judgments <str<strong>on</strong>g>for</str<strong>on</strong>g> each axis of the two comp<strong>on</strong>ent(comp.) space <strong>and</strong> the different difficulties (diff.): easy, medium (med.) <strong>and</strong> hard. .................... 128Figure 7.6. Boxplot of the EEG_W values <str<strong>on</strong>g>for</str<strong>on</strong>g> the three c<strong>on</strong>diti<strong>on</strong>. The red line represent themedian of the EEG_W values, the box the quartile <strong>and</strong> the whiskers the range. NS: n<strong>on</strong>significant. ......................................................................................................................................... 130Figure 7.7. Accuracies of the different classifiers <strong>and</strong> feature selecti<strong>on</strong> metods <strong>on</strong> the peripheralfeatures. .............................................................................................................................................. 133Figure 7.8. Histograms of the number of cross-validati<strong>on</strong> iterati<strong>on</strong>s (over a total of 20) in whichfeatures have been selected by the FCBF, ANOVA <strong>and</strong> SFFS feature selecti<strong>on</strong> algorithms. TheSFFS feature selecti<strong>on</strong> is displayed <str<strong>on</strong>g>for</str<strong>on</strong>g> the DQDA classificati<strong>on</strong>. ................................................ 134Figure 7.9. Accuracies of the different classifiers <strong>and</strong> feature selecti<strong>on</strong> metods <strong>on</strong> the EEGfeatures. .............................................................................................................................................. 135Figure 7.10. Histograms of the number of cross-validati<strong>on</strong> iterati<strong>on</strong>s (over a total of 14) in whichfeatures have been selected by the FCBF, ANOVA <strong>and</strong> SFFS feature selecti<strong>on</strong> algorithms. TheSFFS feature selecti<strong>on</strong> is displayed <str<strong>on</strong>g>for</str<strong>on</strong>g> the DLDA classificati<strong>on</strong>. ................................................. 136Figure 7.11. Classificati<strong>on</strong> accuracy as a functi<strong>on</strong> of the durati<strong>on</strong> of a trial <str<strong>on</strong>g>for</str<strong>on</strong>g> EEG <strong>and</strong> peripheralfeatures. .............................................................................................................................................. 138Figure 7.12.Averages of the normalized GSR <strong>and</strong> HR signals <str<strong>on</strong>g>for</str<strong>on</strong>g> the 5 sec<strong>on</strong>ds following thegame-over triggers. Points that are marked with a ‘+’ corresp<strong>on</strong>ds to the samples that were foundto be significantly different (p-value < 0.1) am<strong>on</strong>g the two c<strong>on</strong>diti<strong>on</strong>s. ‘**’ indicate that p-value


List of tablesList of tablesTable 2.1. Lists of basic emoti<strong>on</strong>s from a biological point of view (from [33]). ............................ 15Table 2.2. List <strong>and</strong> descripti<strong>on</strong> of the different Stimulus Evaluati<strong>on</strong> Checks (SECs) grouped byappraisal objective <strong>and</strong> temporally ordered. ...................................................................................... 22Table 2.3. Comparis<strong>on</strong> of the different methods <str<strong>on</strong>g>for</str<strong>on</strong>g> the m<strong>on</strong>itoring of brain activity. .................. 30Table 2.4. List of publicati<strong>on</strong>s <strong>on</strong> emoti<strong>on</strong> assessment from physiological signals. Signalsacr<strong>on</strong>yms are: Electromyography (EMG), Electrocardiogram (ECG), Galvanic Skin Resp<strong>on</strong>se(GSR), Electroencephalography (EEG), Blood Volume Pulse (BVP). Classificati<strong>on</strong> acr<strong>on</strong>yms are: Sequential Floating Forward Search (SFFS), Linear discriminant analysis (LDA), SupportVector Machine (SVM), Mean Square Error (MSE), Multi Layer Perceptr<strong>on</strong> (MLP), K-NearestNeighbors (KNN), ANalysis Of Variance (ANOVA). ..................................................................... 45Table 3.1. Low <strong>and</strong> high pass cutoff frequencies at -3dB <str<strong>on</strong>g>for</str<strong>on</strong>g> the different filters. ......................... 56Table 3.2. The energy features computed <str<strong>on</strong>g>for</str<strong>on</strong>g> each electrode <strong>and</strong> the associated frequency b<strong>and</strong>s............................................................................................................................................................... 61Table 3.3. The pairs of electrodes used to compute 9 asymmetry scores........................................ 62Table 3.4. The three features computed from the HR....................................................................... 67Table 3.5. The power features computed from the respirati<strong>on</strong> signals <strong>and</strong> being part of thefeature vector. ...................................................................................................................................... 67Table 4.1. A c<strong>on</strong>fusi<strong>on</strong> matrix, P i,j is the percentage of samples bel<strong>on</strong>ging to class i <strong>and</strong>classified as class j . ........................................................................................................................... 75Table 5.1. The 18 features extracted from the peripheral signals. ................................................... 92Table 5.2. p-values of the ANOVA test applied <strong>on</strong> the lateralizati<strong>on</strong> features <str<strong>on</strong>g>for</str<strong>on</strong>g> the two groupsdefined by the IAPS classes (negative vs. positive visual stimuli) <strong>and</strong> <str<strong>on</strong>g>for</str<strong>on</strong>g> each participant. p-values < 0.1 are highlited in gray. ...................................................................................................... 97Table 6.1. The features extracted from the peripheral signals ....................................................... 104Table 6.2. Average c<strong>on</strong>fusi<strong>on</strong> matrices across participants <str<strong>on</strong>g>for</str<strong>on</strong>g> peripheral features <strong>and</strong> differentclassifiers: (a) LDA, (b) Linear SVM, (c) RBF SVM <strong>and</strong> (d) Linear RVM. ................................ 112Table 7.1. The features extracted from the peripheral signals. ...................................................... 126PowfResp179


List of tablesTable 7.2. The baseline subtracti<strong>on</strong> strategy used <str<strong>on</strong>g>for</str<strong>on</strong>g> each feature. -: no subtracti<strong>on</strong> of a baseline.L: last value of the baseline signal subtracted. F: the baseline is computed using the same methodthan <str<strong>on</strong>g>for</str<strong>on</strong>g> feature computati<strong>on</strong>. ........................................................................................................... 127Table 7.3. F-values <strong>and</strong> p-values of the ANOVA tests applied <strong>on</strong> the peripheral features <str<strong>on</strong>g>for</str<strong>on</strong>g> the 3difficulty levels. Only the relevant features are presented (p-value < 0.1). The “Trend of themean” column indicates the differences between two c<strong>on</strong>diti<strong>on</strong>s. For instance indicate asignificant decrease of the variable from the easy to the medium c<strong>on</strong>diti<strong>on</strong> (first ) <strong>and</strong> from themedium to the hard c<strong>on</strong>diti<strong>on</strong> (sec<strong>on</strong>d ), while indicate no significant differences betweenthe easy <strong>and</strong> medium c<strong>on</strong>diti<strong>on</strong> <strong>and</strong> a significant increase to the hard c<strong>on</strong>diti<strong>on</strong>. ........................ 129Table 7.4. List of the relevant EEG features (p-value < 0.1) given by frequency b<strong>and</strong> <strong>and</strong>electrode. ............................................................................................................................................ 130Table 7.5. C<strong>on</strong>fusi<strong>on</strong> matrix <str<strong>on</strong>g>for</str<strong>on</strong>g> the DQDA classifier with FCBF feature selecti<strong>on</strong>.................... 135Table 7.6. C<strong>on</strong>fusi<strong>on</strong> matrix <str<strong>on</strong>g>for</str<strong>on</strong>g> the DLDA classifier with ANOVA feature selecti<strong>on</strong>. .............. 137Table 7.7. C<strong>on</strong>fusi<strong>on</strong> matrix <str<strong>on</strong>g>for</str<strong>on</strong>g> the “Bayes belief integrati<strong>on</strong>” fusi<strong>on</strong>. ........................................ 139180

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!