Skip to content

Latest commit

 

History

History
61 lines (33 loc) · 3.24 KB

README.md

File metadata and controls

61 lines (33 loc) · 3.24 KB

Creative-GAN for Novel Rhythm Generation

RhythmCAN – Generating novel rhythm patterns using GAN with Genre Ambiguity Loss

Can we use AI, more specifically GAN, to originate new electronic dance music genres?

Since the introduction of deep learning, researchers have proposed content generation systems using deep learning and proved that they are competent to generate convincing content and artistic output, including music. However, one can argue that these deep learning-based systems imitate and reproduce the patterns inherent within what humans have created, instead of generating something new and creative.
In this paper, we focus on music generation, especially rhythm patterns of electronic dance music, and discuss if we can use deep learning to generate novel rhythms, interesting patterns not found in the training dataset. We extend the framework of Generative Adversarial Networks(GAN) and encourage it to diverge from the inherent distributions in the dataset by adding additional classifiers to the framework. The paper shows that our proposed GAN can generate rhythm patterns that sound like music rhythms but not belong to any genre in the training dataset.

Project Page

https://cclab.sfc.keio.ac.jp/projects/rhythmcan/

Requirements

Before doing pip install -r requirements.txt, edit requirements.txt depending on your needs please.

  • Python 3.6 or later
  • TensorFlow (tested with 1.15.1)
  • keras
  • matplotlib
  • pyfluidsynth etc.

also, you need to install fluidsynth separately by sudo apt-get install fluidsynth.

Generated Rhythms

You can find generated samples of Conditional-GAN and Creative-GAN in ./audio_samples/ directory.

Or here are soundcloud playlists:
Creative-GAN rhythm patterns
Conditional-GAN rhythm patterns

Training

Playback

We provide an Ableton Live M4L device for playing generated rhythm patterns. Once you train a Creative-GAN (or conditional GAN) model with the scripts in this repository, you can convert the .h5 model into a TFJS model.

$ tensorflowjs_converter --input_format=keras \ /path/to/saved_model /path/to/tfjs_model

The following video demonstrates how the device works and how generated beats sound!

We also provide pre-trained TFJS models here. See M4L-CreativeGAN-Rhythm repository for more details.

Reference

We got the inspiration from the following paper:

Elgammal, Ahmed, Bingchen Liu, Mohamed Elhoseiny, and Marian Mazzone. “CAN: Creative Adversarial Networks, Generating "Art" by Learning About Styles and Deviating from Style Norms,” June 21, 2017. http://arxiv.org/abs/1706.07068.