TY - GEN
T1 - An interactive musical prediction system with mixture density recurrent neural networks
AU - Martin, Charles P.
AU - Torresen, Jim
N1 - Publisher Copyright:
© 2019 Steering Committee of the International Conference on New Interfaces for Musical Expression. All rights reserved.
PY - 2019
Y1 - 2019
N2 - This paper is about creating digital musical instruments where a predictive neural network model is integrated into the interactive system. Rather than predicting symbolic music (e.g., MIDI notes), we suggest that predicting future control data from the user and precise temporal information can lead to new and interesting interactive possibilities. We propose that a mixture density recurrent neural network (MDRNN) is an appropriate model for this task. The predictions can be used to fill-in control data when the user stops performing, or as a kind of filter on the user’s own input. We present an interactive MDRNN prediction server that allows rapid prototyping of new NIMEs featuring predictive musical interaction by recording datasets, training MDRNN models, and experimenting with interaction modes. We illustrate our system with several example NIMEs applying this idea. Our evaluation shows that real-time predictive interaction is viable even on single-board computers and that small models are appropriate for small datasets.
AB - This paper is about creating digital musical instruments where a predictive neural network model is integrated into the interactive system. Rather than predicting symbolic music (e.g., MIDI notes), we suggest that predicting future control data from the user and precise temporal information can lead to new and interesting interactive possibilities. We propose that a mixture density recurrent neural network (MDRNN) is an appropriate model for this task. The predictions can be used to fill-in control data when the user stops performing, or as a kind of filter on the user’s own input. We present an interactive MDRNN prediction server that allows rapid prototyping of new NIMEs featuring predictive musical interaction by recording datasets, training MDRNN models, and experimenting with interaction modes. We illustrate our system with several example NIMEs applying this idea. Our evaluation shows that real-time predictive interaction is viable even on single-board computers and that small models are appropriate for small datasets.
UR - http://www.scopus.com/inward/record.url?scp=85098798525&partnerID=8YFLogxK
M3 - Conference contribution
AN - SCOPUS:85098798525
T3 - Proceedings of the International Conference on New Interfaces for Musical Expression
SP - 260
EP - 265
BT - Proceedings of the International Conference on New Interfaces for Musical Expression
T2 - 19th International conference on New Interfaces for Musical Expression, NIME 2019
Y2 - 3 June 2019 through 6 June 2019
ER -