Just messing around with LSTM neural nets, I had the idea to train it on some guitar tabs and see what happened.
Data
My dataset was a collection of guitar tabs for The Who’s album, Tommy.
Some stats
Text total length: 277,155
Distinct chars : 91
Total sequences : 92,369
So yes, it’s a tiny little dataset.
Training
I’m using TFLearn’s implementation of an LSTM neural net. My setup is as follows:
Input data -> LSTM layer (512 units, returns full sequence (3d tensor)), -> dropout (50%) -> another LSTM layer (512 units, returns only output of last sequence (2d tensor)) -> dropout (50%) -> fully connected layer (softmax activation).
Results
On the first epoch with a temperature of 1.0 (meaning we’re trying to generate novel sequences, less likely to repeat the input data), we get: