MuseGAN is a deep learning research project designed to generate symbolic music using generative adversarial networks. The system focuses specifically on generating multi-track polyphonic music, meaning that it can simultaneously produce multiple instrument parts such as drums, bass, piano, guitar, and strings. Instead of generating raw audio, the model operates on piano-roll representations of music, which encode notes as time-pitch matrices for each instrument track. This representation allows the neural network to capture rhythmic patterns, harmonic relationships, and structural dependencies across instruments. The architecture is based on convolutional GAN models that learn temporal musical structure and inter-track relationships from training data. The project was trained using the Lakh Pianoroll Dataset, a large collection of multitrack musical sequences derived from MIDI files.
Features
- Generative adversarial network architecture for symbolic music generation
- Training pipeline using large pianoroll datasets derived from MIDI music
- Support for multi-track polyphonic composition with several instruments
- Convolutional neural networks that capture temporal musical structure
- Tools for generating music from scratch or conditioning on existing tracks
- Export of generated compositions as pianoroll matrices, images, or MIDI files