19
views
0
recommends
+1 Recommend
0 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      CorrNet: Fine-Grained Emotion Recognition for Video Watching Using Wearable Physiological Sensors

      research-article

      Read this article at

      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          Recognizing user emotions while they watch short-form videos anytime and anywhere is essential for facilitating video content customization and personalization. However, most works either classify a single emotion per video stimuli, or are restricted to static, desktop environments. To address this, we propose a correlation-based emotion recognition algorithm (CorrNet) to recognize the valence and arousal (V-A) of each instance (fine-grained segment of signals) using only wearable, physiological signals (e.g., electrodermal activity, heart rate). CorrNet takes advantage of features both inside each instance (intra-modality features) and between different instances for the same video stimuli (correlation-based features). We first test our approach on an indoor-desktop affect dataset (CASE), and thereafter on an outdoor-mobile affect dataset (MERCA) which we collected using a smart wristband and wearable eyetracker. Results show that for subject-independent binary classification (high-low), CorrNet yields promising recognition accuracies: 76.37 % and 74.03 % for V-A on CASE, and 70.29 % and 68.15 % for V-A on MERCA. Our findings show: (1) instance segment lengths between 1–4 s result in highest recognition accuracies (2) accuracies between laboratory-grade and wearable sensors are comparable, even under low sampling rates (≤64 Hz) (3) large amounts of neutral V-A labels, an artifact of continuous affect annotation, result in varied recognition performance.

          Related collections

          Most cited references123

          • Record: found
          • Abstract: found
          • Article: not found

          Deep learning in neural networks: An overview

          In recent years, deep artificial neural networks (including recurrent ones) have won numerous contests in pattern recognition and machine learning. This historical survey compactly summarizes relevant work, much of it from the previous millennium. Shallow and Deep Learners are distinguished by the depth of their credit assignment paths, which are chains of possibly learnable, causal links between actions and effects. I review deep supervised learning (also recapitulating the history of backpropagation), unsupervised learning, reinforcement learning & evolutionary computation, and indirect search for short programs encoding deep and large networks.
            Bookmark
            • Record: found
            • Abstract: not found
            • Article: not found

            A circumplex model of affect.

              Bookmark
              • Record: found
              • Abstract: found
              • Article: not found

              LSTM: A Search Space Odyssey

              Several variants of the long short-term memory (LSTM) architecture for recurrent neural networks have been proposed since its inception in 1995. In recent years, these networks have become the state-of-the-art models for a variety of machine learning problems. This has led to a renewed interest in understanding the role and utility of various computational components of typical LSTM variants. In this paper, we present the first large-scale analysis of eight LSTM variants on three representative tasks: speech recognition, handwriting recognition, and polyphonic music modeling. The hyperparameters of all LSTM variants for each task were optimized separately using random search, and their importance was assessed using the powerful functional ANalysis Of VAriance framework. In total, we summarize the results of 5400 experimental runs ( ≈ 15 years of CPU time), which makes our study the largest of its kind on LSTM networks. Our results show that none of the variants can improve upon the standard LSTM architecture significantly, and demonstrate the forget gate and the output activation function to be its most critical components. We further observe that the studied hyperparameters are virtually independent and derive guidelines for their efficient adjustment.
                Bookmark

                Author and article information

                Journal
                Sensors (Basel)
                Sensors (Basel)
                sensors
                Sensors (Basel, Switzerland)
                MDPI
                1424-8220
                24 December 2020
                January 2021
                : 21
                : 1
                : 52
                Affiliations
                [1 ]Multimedia Computing Group, Delft University of Technology, 2600AA Delft, The Netherlands; A.Hanjalic@ 123456tudelft.nl
                [2 ]Centrum Wiskunde & Informatica (CWI), 1098XG Amsterdam, The Netherlands; abdallah.el.ali@ 123456cwi.nl
                [3 ]Future Media and Convergence Institute, Xinhuanet & State Key Laboratory of Media Convergence Production Technology and Systems, Xinhua News Agency, Beijing 100000, China; wangchen@ 123456news.cn
                Author notes
                Author information
                https://orcid.org/0000-0001-6293-881X
                https://orcid.org/0000-0002-9954-4088
                https://orcid.org/0000-0002-5771-2549
                https://orcid.org/0000-0003-1752-6837
                Article
                sensors-21-00052
                10.3390/s21010052
                7795677
                33374281
                1d26a795-270f-412a-8a4a-5c541f512de3
                © 2020 by the authors.

                Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license ( http://creativecommons.org/licenses/by/4.0/).

                History
                : 03 December 2020
                : 21 December 2020
                Categories
                Article

                Biomedical engineering
                emotion recognition,video,physiological signals,machine learning
                Biomedical engineering
                emotion recognition, video, physiological signals, machine learning

                Comments

                Comment on this article