In this work, we present DECAF-a multimodal data set for decoding user physiological responses to affective multimedia content. Different from data sets such as DEAP [15] and MAHNOB-HCI [31], DECAF contains (1) brain signals acquired using the Magnetoencephalogram (MEG) sensor, which requires little physical contact with the user's scalp and consequently facilitates naturalistic affective response, and (2) explicit and implicit emotional responses of 30 participants to 40 one-minute music video segments used in [15] and 36 movie clips, thereby enabling comparisons between the EEG versus MEG modalities as well as movie versus music stimuli for affect recognition. In addition to MEG data, DECAF comprises synchronously recorded near-infra-red (NIR) facial videos, horizontal Electrooculogram (hEOG), Electrocardiogram (ECG), and trapezius-Electromyogram (tEMG) peripheral physiological responses. To demonstrate DECAF's utility, we present (i) a detailed analysis of the correlations between participants' self-assessments and their physiological responses and (ii) single-trial classification results for valence, arousal and dominance, with performance evaluation against existing data sets. DECAF also contains time-continuous emotion annotations for movie clips from seven users, which we use to demonstrate dynamic emotion prediction.

DECAF: MEG-based Multimodal Database for Decoding Affective Physiological Responses / Khomami Abadi, Mojtaba; Subramanian, Ramanathan; Kia, Seyed Mostafa; Avesani, Paolo; Patras, Ioannis; Sebe, Niculae. - In: IEEE TRANSACTIONS ON AFFECTIVE COMPUTING. - ISSN 1949-3045. - 2015, 6:3(2015), pp. 209-222. [10.1109/TAFFC.2015.2392932]

DECAF: MEG-based Multimodal Database for Decoding Affective Physiological Responses

Khomami Abadi, Mojtaba;Subramanian, Ramanathan;Kia, Seyed Mostafa;Sebe, Niculae
2015-01-01

Abstract

In this work, we present DECAF-a multimodal data set for decoding user physiological responses to affective multimedia content. Different from data sets such as DEAP [15] and MAHNOB-HCI [31], DECAF contains (1) brain signals acquired using the Magnetoencephalogram (MEG) sensor, which requires little physical contact with the user's scalp and consequently facilitates naturalistic affective response, and (2) explicit and implicit emotional responses of 30 participants to 40 one-minute music video segments used in [15] and 36 movie clips, thereby enabling comparisons between the EEG versus MEG modalities as well as movie versus music stimuli for affect recognition. In addition to MEG data, DECAF comprises synchronously recorded near-infra-red (NIR) facial videos, horizontal Electrooculogram (hEOG), Electrocardiogram (ECG), and trapezius-Electromyogram (tEMG) peripheral physiological responses. To demonstrate DECAF's utility, we present (i) a detailed analysis of the correlations between participants' self-assessments and their physiological responses and (ii) single-trial classification results for valence, arousal and dominance, with performance evaluation against existing data sets. DECAF also contains time-continuous emotion annotations for movie clips from seven users, which we use to demonstrate dynamic emotion prediction.
2015
3
Khomami Abadi, Mojtaba; Subramanian, Ramanathan; Kia, Seyed Mostafa; Avesani, Paolo; Patras, Ioannis; Sebe, Niculae
DECAF: MEG-based Multimodal Database for Decoding Affective Physiological Responses / Khomami Abadi, Mojtaba; Subramanian, Ramanathan; Kia, Seyed Mostafa; Avesani, Paolo; Patras, Ioannis; Sebe, Niculae. - In: IEEE TRANSACTIONS ON AFFECTIVE COMPUTING. - ISSN 1949-3045. - 2015, 6:3(2015), pp. 209-222. [10.1109/TAFFC.2015.2392932]
File in questo prodotto:
File Dimensione Formato  
07010926 Sebe.pdf

Solo gestori archivio

Tipologia: Versione editoriale (Publisher’s layout)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 900.6 kB
Formato Adobe PDF
900.6 kB Adobe PDF   Visualizza/Apri
DECAF-TAFC2015.pdf

accesso aperto

Tipologia: Post-print referato (Refereed author’s manuscript)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 4.7 MB
Formato Adobe PDF
4.7 MB Adobe PDF Visualizza/Apri

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11572/98185
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 248
  • ???jsp.display-item.citation.isi??? 219
  • OpenAlex ND
social impact