Biosignal Datasets for Emotion Recognition

At the HCI Games Group, we love looking at emotion as a core driver of gameplay experience. One common technique used to find out how players experience a game prototype and what affective responses in-game interaction triggers, is to ask players how they feel after playing the game. For this purpose, different affective dimensions like arousal (i.e., the level of excitement), valence (i.e., good or bad) or dominance (i.e., how much the player felt in control) are often used to quantify subjective phenomena. As you can imagine, these types of self-reports are extremely valuable to iteratively improve a game prototype.

Self-Assessment Manikin used to quantify affective dimensions (valence and arousal)

However, one drawback of post-hoc questionnaires is that certain types of emotions are temporary bursts of experience which may fade over time. This becomes a problem if the goal is to investigate affective responses in real-time. To work around this problem, the use of biosignals like brain activity (e.g., through electroencephalography or magnetoencephalography), heart rate (e.g., through electrocardiography), skin conductance level, skin temperature or muscular activity (via electromyograms) has been suggested in the literature [1]. A major focus of the HCI Games Group is in this area.

Repidly is a collaborative tool for biosignal-based gameplay analysis, developed in the HCI Games Group

The main challenge of this biometric approach lies in finding a reliable mapping from physiological patterns to affective states as experienced by the player. The goal is to quantify certain affective dimensions in real-time, without interrupting the game flow to interview the player about their emotional state. Therefore, it is helpful to create automated methods to estimate amplitudes of affective dimensions based on real-time physiological measurements on the fly. However, datasets containing all the information needed to create such automated methods, are costly, hard to find online, and tedious to acquire yourself.

In this blog post, we hope to provide a usable synopsis of freely available datasets that can be used by anyone who is interested in doing research in this area. The four datasets we summarize in this post are MAHNOB-HCI (Soleymani et al.), EMDB (Carvalho et al.), DEAP (Koelstra et al.), and DECAF (Abadi et al.). All these datasets were acquired by presenting multimedia content (i.e., images or videos) to the participants, and recording various physiological responses the participants had to this content. To allow for a mapping from physiological to affective responses, all of the datasets contain subjective self-reports about affective dimensions like arousal, valence, and dominance.


MAHNOB-HCI (Multimodal Database for Affect Recognition and Implicit Tagging)

Authors/Institutions: Soleymani et al. (Imperial College London, Intelligent Behaviour Understanding Group)
Year: 2011
Website: MAHNOB Databases
Publication: A Multimodal Database for Affect Recognition and Implicit Tagging [2]
Participants: 27 (11 male, 16 female)
Stimuli: fragments of movies and pictures
Recorded Signals: 6 face and body cameras (60fps), head-worn microphone (44.1kHz), eye gaze tracker (60Hz), electrocardiogram (ECG), electroencephalogram (32 channels), skin temperature and respiration amplitude (all biosignals at 256Hz)
Subjective scores: arousal, valence, dominance and predictability (both on a scale from 1 to 9)
How to access: Follow the instructions on the request a user account page. In order to receive access to the dataset, you will need to sign an end user license agreement.


EMDB (Emotional Movie Database)

Authors: Carvalho et al.
Year: 2012
Publication: The Emotional Movie Database (EMDB): A Self-Report and Psychophysiological Study [3]
Participants: 113 for self-report ratings of movie clips and 32 for biometrics
Stimuli: 52 movie clips without auditory content from different emotional categories
Recorded Signals: skin conductance level (SCL) and heart rate (HR)
Subjective scores: arousal, valence and dominance (all on a scale from 1 to 9)
How to access: Send a request for the database to EMDB@psi.uminho.ptand follow the instructions in the response. When submitting a request to use the database, you will be asked to confirm that it will be used solely for non-profit scientific research and that the database will not be reproduced or broadcast in violation of international copyright laws.


DEAP (Database for Emotion Analysis using Physiological Signals)

Authors/Institutions: Koelstra et al. (peta media, Queen Mary University of London, University of Twente, University of Geneva, EPFL)
Year: 2012
Website: DEAP Dataset
Publication: DEAP: A Database for Emotion Analysis ;Using Physiological Signals[4]
Participants: 32
Stimuli: 40x 1-minute long excerpts from music videos from last.fm (retrieved using affective tags, video highlight detection and an online assessment tool)
Recorded Signals: electroencephalogram (32 channels at 512Hz), skin conductance level (SCL), respiration amplitude, skin temperature, electrocardiogram, blood volume by plethysmograph, electromyograms of Zygomaticus and Trapezius muscles (EMGs), electrooculogram (EOG), face video (for 22 participants)
Subjective scores: arousal, valence, like/dislike, dominance (all on a scale from 1 to 9), familiarity (on a scale from 1 to 5)
How to access: Download, print and sign the end user license agreement and send a signed copy to the authors. You will then receive a password to be entered when you click on the download links, as described here


DECAF (Multimodal Dataset for Decoding Affective Physiological Responses)

Authors/Institutions: Abadi et al. (University of Trento, FBK, ADSC, CiMeC, Semantics & Knowledge Innovation Lab, Telecom Italia)
Year: 2015
Website: MHUG
Publication: DECAF: MEG-Based Multimodal Database for Decoding Affective Physiological Responses [5]
Participants: 30
Stimuli: 40x 1-minute long excerpts from music videos (same as in the DEAP dataset), 36x movie clips
Recorded Signals: magnetoencephalogram (MEG), horizontal electrooculogram (hEOG), electrocardiogram (ECG), electromyogram of the Trapezius muscle (tEMG), near-infrared face video
Subjective scores: arousal, valence, dominance (all on a scale from 1 to 9), time-continuous emotion annotations for movie clips (from 7 experts)
How to access: Download, print and sign the end user license agreement and send a signed copy to the authors. You will then receive a password to be entered when you click on the download links, as described here


As you can see, not all of the datasets use the same stimuli or collect the same types of physiological and self-reported data. It is interesting to note, that the common denominator for self-reported data are the three affective scores (arousal, valence, and dominance) collected on a discrete scale from 1 to 9. As for physiological recordings, heart rate is the only metric that can be found in all four datasets. EEG, skin temperature and respiration amplitude are collected in MAHNOB-HCI and DEAP. EMG of the Trapezius muscle and EOG are collected in both DEAP and DECAF. Skin conductance levels are available in DEAP and EMDB. In contrast, face videos are recorded for MANHOB-HCI, DEAP and DECAF. Finally, DECAF is the only dataset providing MEG brain activity data. Luckily however, as DECAF uses the same music video stimuli as DEAP. Correlations between the EEG brain responses found in DEAP and the MEG brain responses found in DECAF, can be investigated.

We really hope to have provided a fair and useful overview on these four datasets and look forward to seeing more research being done in the area of biosignal-based affect recognition in the future — as we are really excited about this topic.


Contact

If you are interested and have some more questions, I would love you to get in touch — either via Twitter: @HardyShakerman or email: mikeschaekermann@gmail.com. Cheers, Mike Schaekermann


Acknowledgements

We would love to thank the creators of these wonderful datasets for investing an incredible amount of time and work into data collection and for making these invaluable datasets freely available for research.



References

[1] Nacke, L., & Lindley, C. A. (2008). Flow and immersion in first-person shooters: measuring the player’s gameplay experience. In Proceedings of the 2008 Conference on Future Play: Research, Play, Share (pp. 81–88). New York, NY, USA: ACM. http://doi.org/10.1145/1496984.1496998

[2] Soleymani, M., Lichtenauer, J., Pun, T., & Pantic, M. (2012). A Multimodal Database for Affect Recognition and Implicit Tagging. IEEE Transactions on Affective Computing, 3(1), 42–55. http://doi.org/10.1109/T-AFFC.2011.25

[3] Carvalho, S., Leite, J., Galdo-Álvarez, S., & Gonçalves, Ó. F. (2012). The Emotional Movie Database (EMDB): A Self-Report and Psychophysiological Study. Applied Psychophysiology and Biofeedback, 37(4), 279–294. http://doi.org/10.1007/s10484-012-9201-6

[4] Koelstra, S. (2012). Deap: A database for emotion analysis; using physiological signals. Affective Computing, 3(1), 18–31. http://doi.org/10.1109/T-AFFC.2011.15

[5] Abadi, M. K., Subramanian, R., Kia, S. M., Avesani, P., Patras, I., & Sebe, N. (2015). DECAF: MEG-Based Multimodal Database for Decoding Affective Physiological Responses. IEEE Transactions on Affective Computing, 6(3), 209–222. http://doi.org/10.1109/TAFFC.2015.2392932

Read next

HCI Games Group at CHI PLAY 2024

Our HCI Games Group members are set to present their research at CHI PLAY 2024, held from October 14-17, 2024, in Tampere, Finland. This year, Lennart Nacke, Hilda Hadan, Sabrina Alicia Sgandurra, Sebastian Cmentowski, Reza Hadi Mogavi, Derrick M. Wang,