A First Look at Music Composition using LSTM Recurrent Neural Networks

Slides:



Advertisements
Similar presentations
Dougal Sutherland, 9/25/13.
Advertisements

Memristor in Learning Neural Networks
Neural networks Introduction Fitting neural networks
1 Machine Learning: Lecture 4 Artificial Neural Networks (Based on Chapter 4 of Mitchell T.., Machine Learning, 1997)
NEURAL NETWORKS Perceptron
1 Neural networks. Neural networks are made up of many artificial neurons. Each input into the neuron has its own weight associated with it illustrated.
Artificial Intelligence 13. Multi-Layer ANNs Course V231 Department of Computing Imperial College © Simon Colton.
CSC321: 2011 Introduction to Neural Networks and Machine Learning Lecture 7: Learning in recurrent networks Geoffrey Hinton.
Machine Learning Neural Networks
Decision Support Systems
Simple Neural Nets For Pattern Classification
Branch Prediction with Neural- Networks: Hidden Layers and Recurrent Connections Andrew Smith CSE Dept. June 10, 2004.
Introduction to Recurrent neural networks (RNN), Long short-term memory (LSTM) Wenjie Pei In this coffee talk, I would like to present you some basic.
Artificial Neural Networks
Artificial Neural Network
Neural Networks. Background - Neural Networks can be : Biological - Biological models Artificial - Artificial models - Desire to produce artificial systems.
Artificial Neural Networks (ANN). Output Y is 1 if at least two of the three inputs are equal to 1.
Deep Learning Neural Network with Memory (1)
NEURAL NETWORKS FOR DATA MINING
Artificial Intelligence Methods Neural Networks Lecture 4 Rakesh K. Bissoondeeal Rakesh K. Bissoondeeal.
Artificial Neural Networks. The Brain How do brains work? How do human brains differ from that of other animals? Can we base models of artificial intelligence.
ARTIFICIAL NEURAL NETWORKS. Overview EdGeneral concepts Areej:Learning and Training Wesley:Limitations and optimization of ANNs Cora:Applications and.
Chapter 18 Connectionist Models
Neural Networks Vladimir Pleskonjić 3188/ /20 Vladimir Pleskonjić General Feedforward neural networks Inputs are numeric features Outputs are in.
Supervised Sequence Labelling with Recurrent Neural Networks PRESENTED BY: KUNAL PARMAR UHID:
Chapter 6 Neural Network.
Neural Networks Lecture 11: Learning in recurrent networks Geoffrey Hinton.
Predicting the dropouts rate of online course using LSTM method
NOTE: To change the image on this slide, select the picture and delete it. Then click the Pictures icon in the placeholder to insert your own image. SHOW.
Business Intelligence and Decision Support Systems (9 th Ed., Prentice Hall) Chapter 6: Artificial Neural Networks for Data Mining.
Combining Models Foundations of Algorithms and Machine Learning (CS60020), IIT KGP, 2017: Indrajit Bhattacharya.
Today’s Lecture Neural networks Training
Neural networks.
Deep Learning RUSSIR 2017 – Day 3
Convolutional Sequence to Sequence Learning
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 2, 2017.
RNNs: An example applied to the prediction task
CS 388: Natural Language Processing: LSTM Recurrent Neural Networks
Deep Learning Amin Sobhani.
Learning in Neural Networks
Advanced information retreival
Recursive Neural Networks
Recurrent Neural Networks for Natural Language Processing
COMP24111: Machine Learning and Optimisation
ICS 491 Big Data Analytics Fall 2017 Deep Learning
Intelligent Information System Lab
Different Units Ramakrishna Vedantam.
CSE P573 Applications of Artificial Intelligence Neural Networks
Attention Is All You Need
Generalization ..
CSE 473 Introduction to Artificial Intelligence Neural Networks
RNNs: Going Beyond the SRN in Language Prediction
Artificial Neural Networks for Pattern Recognition
A critical review of RNN for sequence learning Zachary C
Grid Long Short-Term Memory
Image Captions With Deep Learning Yulia Kogan & Ron Shiff
Recurrent Neural Networks
Learning linguistic structure with simple and more complex recurrent neural networks Psychology February 8, 2018.
CSE 573 Introduction to Artificial Intelligence Neural Networks
The use of Neural Networks to schedule flow-shop with dynamic job arrival ‘A Multi-Neural Network Learning for lot Sizing and Sequencing on a Flow-Shop’
Long Short Term Memory within Recurrent Neural Networks
Other Classification Models: Recurrent Neural Network (RNN)
Recurrent Encoder-Decoder Networks for Time-Varying Dense Predictions
RNNs: Going Beyond the SRN in Language Prediction
实习生汇报 ——北邮 张安迪.
Deep Learning Authors: Yann LeCun, Yoshua Bengio, Geoffrey Hinton
CSC321: Neural Networks Lecture 11: Learning in recurrent networks
Recurrent Neural Networks
Random Neural Network Texture Model
LHC beam mode classification
Presentation transcript:

A First Look at Music Composition using LSTM Recurrent Neural Networks Written by: Douglas Eck, Jurgen Schmidhuber Presented by: Danit Itzkovich, Mor Yemini

Recurrent Neural Network (RNN) Designed to learn sequential or time varying patterns A neural network with feedback (closed loop) connections

Recurrent Neural Network (RNN)

Long Short Term Memory (LSTM)

Long Short Term Memory (LSTM)

Long Short Term Memory (LSTM)

Music Composition using LSTM example

compose music with computers Music generation can be formulated as a time-series prediction problem: the note played at each time can be regarded as a prediction given the notes played before. Music is a complex structure that has both short-term and long-term dependency just as a language model makes recurrent neural network (RNN) an ideal structure for this task.

compose music with RNN Note by note approach : Single step predictor. Predicts note at time t+1 using notes at time t. At test time, one input note is needed then the network generates a novel composition.

The problem Vanishing gradient The network is impossible of learning long term dependencies, which are critical in music composition. In the case of music, long-term dependencies are at the heart of what defines a particular style.

Compose music with LSTM The solution Compose music with LSTM

A First Look at Music Composition using LSTM Recurrent Neural Networks By: Douglas Eck, Jurgen Schmidhuber

Training data A range of 12 notes were possible for chords. A range of 13 notes were possible for melodies. A single song length was 96 notes. The chords used did not vary from song to song. Only quarter notes were used. No rests were used.

Training data

Data representation One input/target unit per note, with 1.0 representing on and 0.0 representing off. The Network input represented the whole song.

Experiment 1 – learning Chords The goal: Determine whether LSTM can learn to reproduce a musical chord structure with the absence of melody. Predict the probability for a given note to be on or off.

Experiment 1 – learning Chords Network topology and experimental parameters: Four cell blocks containing two cells each were fully connected to each other and to the input layer. The output layer is fully connected to all cells and to the input layer. Forget gate, input gate and output gate biases for the four blocks were set at -0.5, -1.0, -1.5 and -2.0.

Experiment 1 – learning Chords Network topology and experimental parameters: Learning rate was set at 0.00001. Momentum rate was set at 0.9. Reset weight: error burn existing weights reset the input pattern and clear partial derivatives, activations and cell states. Output activation function was sigmoid.

Experiment 1 – learning Chords Training: This task is similar to multi-label classification problem The network was trained using cross-entropy function. The error function Target value Output activation predicted value Training was stopped after the network successfully predicted the entire chord sequence.

Experiment 1 – learning Chords Testing: The testing phase is also the composition phase, where the predictions could be recorded to form new songs. The network was tested by starting it with the inputs from the first time step and then using network predictions for ensuing time steps. Chord notes were predicted using a decision threshold of 0.5.

Experiment 1 – learning Chords Results: LSTM easily handled the task under a wide range of learning rates and momentum rates.

Experiment 2 – Learning Melody and Chords The goal: Determine whether LSTM could learn chord structure and melody structure and then use that structure to advantage when composing new songs. Learn melody and chords simultaneously. Predict the probability for a given note to be on or off.

Experiment 2 – Learning Melody and Chords Network topology and experimental parameters: Some cell blocks processed chord information while other cell blocks processed melody information. Eight cell blocks containing two cells each are used: Four of the cell blocks are fully connected to the input units for chords. Four cell blocks are fully connected to the input units for melody. The chord cell blocks have recurrent connections to themselves and to the melody cell blocks. Melody cell blocks are only recurrently connected to melody cell blocks.

Experiment 2 – Learning Melody and Chords Network topology and experimental parameters: Output units for chords are fully connected to cell blocks for chords and to input units for chords. Output units for melody are fully connected to cell blocks for melody and to input units for melody. Forget gate, input gate and output gate biases for the four blocks were set at -0.5, -1.0, -1.5 and -2.0. Learning rate was set at 0.00001. Momentum rate was set at 0.9.

Experiment 2 – Learning Melody and Chords Training: For chords – the same training algorithm as Experiment 1 was used. For melodies – melody output activations summed to 1.0 to choose a single note at any given time step. The network was trained until it had learned the chord structure and until objective error had reached a plateau.

Experiment 2 – Learning Melody and Chords Testing: The network was allowed to freely compose music. Music was composed by providing a single note or series of notes (up to 24) from the training set as input.

Experiment 2 – Learning Melody and Chords Results: In all cases the network succeeded in reproducing chord structure while in parallel improvising new melodies. The network compositions were better sounding than a random walk across the pentatonic scale

conclusions LSTM can capture both the local structure of melody and the long-term structure of a of a musical style The first experiment verified that LSTM was not relying on regularities in the melody to learn the chord structure. The second experiment explored the ability for LSTM to generate new instances of a musical form.

discussion There was no variety in the underlying chord structure. This experiment is suitable for jazz music, in which chord changes are almost always provided separately from melodies. LSTM provided the best results compared to other RNN techniques.

Questions?