Blog
About

30
views
0
recommends
+1 Recommend
1 collections
    0
    shares
      • Record: found
      • Abstract: found
      • Article: found
      Is Open Access

      Generating Time: Rhythmic Perception, Prediction and Production with Recurrent Neural Networks

      , ,

      Journal of Creative Music Systems

      University of Huddersfield Press

      Read this article at

      ScienceOpenPublisher
      Bookmark
          There is no author summary for this article yet. Authors can add summaries to their articles on ScienceOpen to make them more accessible to a non-specialist audience.

          Abstract

          In the quest for a convincing musical agent that performs in real time alongside human performers, the issues surrounding expressively timed rhythm must be addressed. Current beat tracking methods are not sufficient to follow rhythms automatically when dealing with varying tempo and expressive timing. In the generation of rhythm, some existing interactive systems ignore the pulse entirely, or fix a tempo after some time spent listening to input. Since music unfolds in time, we take the view that musical timing needs to be at the core of a music generation system. Our research explores a connectionist machine learning approach to expressive rhythm generation, based on cognitive and neurological models. Two neural network models are combined within one integrated system. A Gradient Frequency Neural Network (GFNN) models the perception of periodicities by resonating nonlinearly with the musical input, creating a hierarchy of strong and weak oscillations that relate to the metrical structure. A Long Short-term Memory Recurrent Neural Network (LSTM) models longer-term temporal relations based on the GFNN output. The output of the system is a prediction of when in time the next rhythmic event is likely to occur. These predictions can be used to produce new rhythms, forming a generative model. We have trained the system on a dataset of expressively performed piano solos and evaluated its ability to accurately predict rhythmic events. Based on the encouraging results, we conclude that the GFNN-LSTM model has great potential to add the ability to follow and generate expressive rhythmic structures to real-time interactive systems.

          Related collections

          Author and article information

          Journal
          Journal of Creative Music Systems
          University of Huddersfield Press
          2399-7656
          March 1 2017
          March 1 2017
          : 1
          : 2
          Article
          10.5920/JCMS.2017.04
          © 2017

          JCMS provides free online access to its content. The work is published under a Creative Commons license, which provides unrestricted use, distribution and reproduction in any medium, provided the original work is properly cited. To view the licenses, visit http://creativecommons.org/licenses/

          Product
          Self URI (journal page): https://www.jcms.org.uk/
          Self URI (article page): https://www.jcms.org.uk/article/id/511/

          Software engineering, Music, Computer science

          Comments

          Comment on this article