+1 Recommend
1 collections

      Celebrating 65 years of The Computer Journal - free-to-read perspectives - bcs.org/tcj65

      • Record: found
      • Abstract: found
      • Conference Proceedings: found
      Is Open Access

      An Interactive Music Playlist Generator that Responds to User Emotion and Context

      1 , 1 , 1
      Electronic Visualisation and the Arts (EVA)
      Electronic Visualisation and the Arts
      12 - 14 July 2016


            This paper aims to demonstrate the mechanisms of a music recommendation system, and accompanying graphical user interface (GUI), that is capable of generating a playlist of songs based upon an individual’s emotion or context. This interactive music playlist generator has been designed as part of a broader system, Intended for mobile devices, which aims to suggest music based upon ‘how the user is feeling’ and ‘what the user is doing’ by evaluating real-time physiological and contextual sensory data using machine learning technologies. For instance, heart rate and skin temperature in conjunction with ambient light, temperature and global positioning satellite (GPS) could be used to a degree to infer one’s current situation and corresponding mood. At present, this interactive music playlist generator has the ability to conceptually demonstrate how a playlist can be formed in accordance with such physiological and contextual parameters. In particular, the affective aspect of the interface is visually represented as a two-dimensional arousal-valence space based upon Russell’s circumplex model of affect (1980).Context refers to environmental, locomotion and activity concepts, and are visually represented in the interface as sliders. These affective and contextual components are discussed in more detail next in Sections 2 and 3, respectively. Section 4 will demonstrate how an affective and contextual music playlist can be formed by interacting with the GUI parameters. For a comprehensive discussion in terms of the development of this research, refer to (Griffiths et al. 2013a, 2013b, 2015). Moreover, refer to Teng et al. (2013) and Yang et al. (2008) for related work in these broader research areas.


            Author and article information

            July 2016
            July 2016
            : 275-276
            [1 ] Glyndŵr University

            Mold Road

            Wrexham LL11 2AW


            This work is licensed under a Creative Commons Attribution 4.0 Unported License. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/

            Electronic Visualisation and the Arts
            London, UK
            12 - 14 July 2016
            Electronic Workshops in Computing (eWiC)
            Electronic Visualisation and the Arts

            1477-9358 BCS Learning & Development

            Self URI (article page): https://www.scienceopen.com/hosted-document?doi=10.14236/ewic/EVA2016.53
            Self URI (journal page): https://ewic.bcs.org/
            Electronic Workshops in Computing

            Applied computer science,Computer science,Security & Cryptology,Graphics & Multimedia design,General computer science,Human-computer-interaction


            1. 2013a Automatic Music Playlist Generation Using Affective Computing Technologies ITA13 September 10-13 177 183

            2. 2013b A Discussion of Musical Features For Automatic Music Playlist Generation Using Affective Technologies AM13 September 16-20

            3. 2015 A Self-Report Study Which Gauges Perceived and Induced Emotion With Music ITA15 September 711 236 241

            4. 1980 A Circumplex Model of Affect. Journal of Personality and Social Psychology 39 6 1161 1178

            5. 2013 A Large In-Situ Dataset For Context-Aware Music Recommendation For Smartphones ICMEW July 15-19 1 4

            6. 2008 Mr. Emo: Music Retrieval in the Emotion Plane MM ‘08 October 26 1003 1004


            Comment on this article