human language technologies data collections & studies wp4- emotions: faces
Post on 31-Mar-2015
219 Views
Preview:
TRANSCRIPT
humanlanguage
technologies
Data Collections & Studies
WP4- Emotions: Faces
Data Collections & StudiesWP4 – Emotions: Faces
Collection and annotation ofaudio-visual databases
extensive data collection, both at KTH and ISTC/IRST– using opto-electronic systems
reflective markers placed on the subject’s face capturing of dynamics of emotional facial expressions with very high precision.
eliciting technique: using movies to elicit facial expressions denoting emotions on watching subjects attempted – not promising
extraction technique: extract expressive behaviour directly from movies and television talk-shows attempted – not promising
Data Collections & StudiesWP4 – Emotions: Faces
KTH - first year
DATABASE 1• 15 emotions and attitudes were recorded (acted)
anger, fear, surprise, sadness, disgust, happiness, worry, satisfaction, insecurity, confidence, questioning, encouragment, doubt, confirmation and neutral
• Semantically neutral utterances, 9 utterances per expressionDATABASE 2• 6 emotional states
confident, confirming, questioning, insecure, happy, neutral• VCV & VCCV nonsense words• CVC nonsense words• Short sentences• Common ITA-SWE set (abba, adda, alla, anna, avva) DATABASE 3• Spontaneous dialogue
Data Collections & StudiesWP4 – Emotions: Faces
Collection of audio-visual databases: interactive dialogues (KTH)
Eliciting technique: information seeking scenario
Focus on the speaker who has the role of information giver
The speaker whose facial and head motion is to be recorded seats facing 4 infrared cameras, a digital video-camera,a microphone and his/her interlocutor.
Data Collections & StudiesWP4 – Emotions: Faces
ISTC & IRST - first year
6 emotional states (Ekman’s set) + Neutral– Anger, Disgust, Fear, Happiness, Sadness, Surprise – 3 intensities (Low, Medium, High)
“isolated” emotional expressions VCV nonsense words (aba, ada, aLA, adZa, ala, ana, ava)
– good phonetic coverage of Italian Long sentence (“Il fabbro lavora con forza usando il martello e la
tenaglia” – “the smith works with strength using the hammer and the pincer”)
common ITA-SWE set (VCCV nonsense words: abba, adda, alla, anna, avva)
“concatenated” emotional expressions VCV nonsense words, in pairs, with different emotions e.g. (aba)Neutral – (aba)Happy
Data Collections & StudiesWP4 – Emotions: Faces
Results
ISTC/IRST: 1573 recordings – 798 single emotional expressions (7 emotional states, 3 intensities –
L, M, H)– 672 concatenated emotional expressions (in pairs, 3 emotional states
- Anger, Happy, Neutral - medium intensity)– 57 long sentences (7 emotional states, 3 intensities)– 46 instances of the common ITA-SWE set (3 emotional states,
medium intensity)
KTH: 1749 recordings (database 2)– 828 VCV words (138 x 6 emotional states)– 246 CVC words (41 x 6 emotional states)– 645 sentences (270 neutral + 75 x 5 emotional states)– 30 instances of the common ITA-SWE set
Total: 3322 recordings
Data Collections & StudiesWP4 – Emotions: Faces
Qualisys recordings: Swedish db – 2nd year
75 sentences with Ekman’s 6 basic emotions + neutral
Dialogues to analyze communicative facial expressions:–10 short dialogues in a travel agency scenario
15 sentences uttered with a focussed word, with the 6 expressions used in corpus 2 + angerExample: Båten seglade förbi
Båten seglade förbiBåten seglade förbi
Data Collections & StudiesWP4 – Emotions: Faces
Short videos containing acted kinetic facial expressions (video length: 4-27 secs.)
8 professional actors (4 male and 4 female).
Each actor played Ekman’s set of six emotions
(happy, sad, angry, surprised, disgusted, afraid) +
neutral
Actors were asked to play each emotion on three
intensity levels (Low -Medium – High)
Total: 1008 short videos (= ~ 2h
50’)
Audio-Visual Italian Database – 2nd year (IRST) : A database of human facial expressions (1)
See Poster!!!
Data Collections & StudiesWP4 – Emotions: Faces
A database of human facial expressions (2)
Facial expressions recorded in two conditions:
1. “utterance” condition: actors played emotions while uttering a phonetically rich and visemically balanced sentence.
2. “non utterance” condition: actors played emotions without
pronouncing any sentence. Both video and audio signals were
recorded. After collecting the corpus: Data Selection
– Validation of the emotions played– Video selection based on the accordance among judges.
“In quella piccola stanza vuota c’era pero’ soltanto una sveglia”, <FEAR>, <HIGH>
See Poster!!!
<DISGUST>, <HIGH>
Data Collections & StudiesWP4 – Emotions: Faces
Annotation of audio-visual databases: interactive dialogues
ANVIL: tool for the analysis of digitized audio-visual data Orthographic transcription of the dialogue Annotation of the facial expressions related to emotions and
of the communicative gestures (turn-taking, feedback and so on)
The annotation is performed on a freely definable multi-layered annotation scheme, created ad hoc for the specific purposes.
These levels go from a less detailed to a more detailed analysis
Annotation is performed on several main tracks, which are displayed, on the screen in alignment with the video and audio data
Data Collections & StudiesWP4 – Emotions: Faces
Annotation (cont’d)
glad
Data Collections & StudiesWP4 – Emotions: Faces
Evaluation Studies (IRST)
Experiment 1: Comparison of emotion recognition rates from natural (actor) videos with different types of synthetic (synthetic face) videos, in different animation conditions
[reference person: Fabio Pianesi – pianesi@itc.it]
Experiment 2: Cross-cultural comparison of emotion recognition rates from Italian and Swedish natural and synthetic videos [reference person: Fabio Pianesi – pianesi@itc.it]
Experiment 3: as for Experiment 1 but using – three regions of the face– only one animation condition (script based)
[reference person: Michela Prete – prete@itc.it]
Data Collections & StudiesWP4 – Emotions: Faces
Papers on Evaluation Studies
J. Beskow, L. Cerrato, P. Cosi, E. Costantini, M. Nordstrand, F. Pianesi, M. Prete, G. Svanfeldt, "Preliminary Cross-cultural Evaluation of Expressiveness in Synthetic Faces". In E. André, L. Dybkiaer, W. Minker, P. Heisterkamp (eds.) "Affective Dialogue Systems", ADS '04, Springer Verlag. Berlin, 2004.
E. Costantini, F. Pianesi, P. Cosi, "Evaluation of Synthetic Faces: Human Recognition of Emotional Facial Displays ". In E. André, L. Dybkiaer, W. Minker, P. Heisterkamp (eds.) "Affective Dialogue Systems". Springer Verlag, Berlin, 2004
E. Costantini, F. Pianesi, M. Prete "Recognising Emotions in Human and Synthetic Faces: The Role of the Upper and Lower Parts of the Face". To appear in Proceedings of IUI 2005: International Conference on Intelligent User Interfaces. San Diego, California, 2005.
top related