The use of AI models for music generation receives an important attention from scientific communities. Different architectures of deep learning neural networks have been applied for this specific task, such as Recurrent Neural Networks (RNN), Generative Adversarial Networks (GAN), Autoencoders, Variational Autoencoders (VAE) and Transformers. One of the important aspects of the generation process is the possibility to control the output by providing the input parameters, and a conditional generation was widely used in a computer vision domain to meet this need. In a present research we adopt the principles of conditional generation using GAN architecture and convolutions, applying them to a temporal domain, resulting in building a conditional GAN for diatonic harmonic sequences generation. The model is further used as a core feature of the VR module for computer-aided composition from “Graphs in harmony learning” VR project, where the sequence generation is conditioned by the user's input and the result is mapped to 3D representations of the generated chords.
M. Cuthbert and C. Ariza (2010) "music21: A Toolkit for Computer-Aided Musicology and Symbolic Music Data," International Conference on Music Information Retrieval, pp. 637–642.
H.W. Dong, W.Y. Hsiao, L.C. Yang and Y.H. Yang, 2018, April. Musegan: Multi-track sequential generative adversarial networks for symbolic music generation and accompaniment. AAAI Conference on Artificial Intelligence (Vol. 32, No. 1).
D. Eck and J. Schmidhuber (2002). Finding Temporal Structure in Music: Blues Improvisation with LSTM Recurrent Networks 12th IEEE Workshop on Neural Networks for Signal Processing.
C.Z.A. Huang, A. Vaswani, J. Uszkoreit, N. Shazeer, I. Simon, C. Hawthorne, A.M. Dai, M.D. Hoffman, M. Dinculescu and D. Eck, 2018. Music transformer. arXiv preprint arXiv:1809.04281.
K. Kritsis, T. Kylafi, M. Kaliakatsos-Papakostas, A. Pikrakis and V. Katsouros (2020). On the Adaptability of Recurrent Neural Networks for Real-Time Jazz Improvisation Accompaniment. Frontiers in Artificial Intelligence, 3, p.113.
F. Liang (2016). Bachbot: Automatic Composition in the s Style of Bach Chorales. M.Phil thesis. University of Cambridge.
A. Shvets (2019). Structural harmony method in the context of deep learning on example of music by Valentyn Sylvestrov and Philipp Glass. In: J. Weinel, J.P. Bowen, G. Diprose, and N. Lambert (eds), EVA London 2019 (Electronic Visualisation and the Arts), London, UK, 10–14 July 2019, 318–320. BCS, London. [Cross Ref]
A. Shvets, S. Darkazanli (2020). Graphs in harmony learning: AI assisted VR application. In: J. Weinel, J.P. Bowen, G. Diprose, and N. Lambert (eds), EVA London 2020 (Electronic Visualisation and the Arts), London, UK, 6–9 July 2020, 104–105. BCS, London. [Cross Ref]
R. Vohra, K. Goel and J. Sahoo (2015). Modelling Temporal Dependencies in Data Using a DBNLSTM. IEEE International Conference on Data Science and Advanced Analytics.
L. C. Yang, S. Y. Chou, & Y. H. Yang (2017). MidiNet: A convolutional generative adversarial network for symbolic-domain music generation. arXiv preprint arXiv:1703.10847.
L. Yu, W. Zhang, J. Wang and Y. Yu (2017). Seqgan: Sequence generative adversarial nets with policy gradient. AAAI conference on artificial intelligence (Vol. 31, No. 1).