5
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: not found
      • Article: not found

      DECAF: MEG-Based Multimodal Database for Decoding Affective Physiological Responses

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Related collections

          Most cited references23

          • Record: found
          • Abstract: not found
          • Article: not found

          Detecting outliers: Do not use standard deviation around the mean, use absolute deviation around the median

            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            A Multimodal Database for Affect Recognition and Implicit Tagging

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              Emotion recognition based on physiological changes in music listening.

              Little attention has been paid so far to physiological signals for emotion recognition compared to audiovisual emotion channels such as facial expression or speech. This paper investigates the potential of physiological signals as reliable channels for emotion recognition. All essential stages of an automatic recognition system are discussed, from the recording of a physiological dataset to a feature-based multiclass classification. In order to collect a physiological dataset from multiple subjects over many weeks, we used a musical induction method which spontaneously leads subjects to real emotional states, without any deliberate lab setting. Four-channel biosensors were used to measure electromyogram, electrocardiogram, skin conductivity and respiration changes. A wide range of physiological features from various analysis domains, including time/frequency, entropy, geometric analysis, subband spectra, multiscale entropy, etc., is proposed in order to find the best emotion-relevant features and to correlate them with emotional states. The best features extracted are specified in detail and their effectiveness is proven by classification results. Classification of four musical emotions (positive/high arousal, negative/high arousal, negative/low arousal, positive/low arousal) is performed by using an extended linear discriminant analysis (pLDA). Furthermore, by exploiting a dichotomic property of the 2D emotion model, we develop a novel scheme of emotion-specific multilevel dichotomous classification (EMDC) and compare its performance with direct multiclass classification using the pLDA. Improved recognition accuracy of 95\% and 70\% for subject-dependent and subject-independent classification, respectively, is achieved by using the EMDC scheme.
                Bookmark

                Author and article information

                Journal
                IEEE Transactions on Affective Computing
                IEEE Trans. Affective Comput.
                Institute of Electrical and Electronics Engineers (IEEE)
                1949-3045
                July 1 2015
                July 1 2015
                : 6
                : 3
                : 209-222
                Article
                10.1109/TAFFC.2015.2392932
                099d2664-6495-496f-8d32-8ca3bc6b44b6
                © 2015
                History

                Comments

                Comment on this article