GithubHelp home page GithubHelp logo

michalliu / deep-music-visualizer Goto Github PK

View Code? Open in Web Editor NEW

This project forked from msieg/deep-music-visualizer

0.0 0.0 0.0 11.19 MB

The Deep Visualizer uses BigGAN (Brock et al., 2018) to visualize music.

Python 100.00%

deep-music-visualizer's Introduction

Deep Music Visualizer

The Deep Music Visualizer uses BigGAN (Brock et al., 2018), a generative neural network, to visualize music. Like this:

Alt text

More examples: https://www.instagram.com/deep_music_visualizer/

Installation

This repo has been tested on Python3

Assuming you have python installed, open terminal and run these commands:

git clone https://github.com/msieg/deep-music-visualizer.git
cd deep-music-visualizer
pip install -r requirements.txt

If you are on linux, you may also need to run:

apt-get update
apt-get install ffmpeg
apt-get install libsndfile1

How to run

All features of the visualizer are available as input parameters. Each parameter is described below.

song

Audio file of type mp3, wav, or ogg.

This is the only required argument!

Example:

python visualize.py --song beethoven.mp3

resolution

128, 256, or 512

Default: 512

If you are running on a CPU (if you're not sure, you are on a CPU), you might want to use a lower resolution or else the code will take a very long time to run. At 512X512, it will take ~7 HOURS to generate 1 minute of video on a standard desktop computer (assuming all other parameters are default). At 128x128, it would take ~25 minutes. To speed up runtime, you can decrease the resolution or increase the frame_length. To dramatically speed up runtime and generate higher quality videos, use a resolution of 512 on a GPU on a google cloud virtual machine.

Example:

python visualize.py --song beethoven.mp3 --resolution 128

duration

Duration of the video output in seconds. It can be useful to generate shorter videos while you are tweaking the other visualizer parameters. Once you find your preferred parameters, remove the duration argument and set use_previous_vectors to 1 to generate the same video but for the full duration of the song.

Default: Full length of the audio

Example:

python visualize.py --song beethoven.mp3 --duration 30

pitch sensitivity

The pitch sensitivity controls how rapidly the class vector (thematic content of the video) will react to changes in pitch. The higher the number, the higher the sensitivity.

Range: 1 – 299

Recommended range: 200 – 295

Default: 220

Example:

python visualize.py --song beethoven.mp3 --pitch_sensitivity 280

tempo sensitivity

The tempo sensitivity controls how rapidly the noise vector (i.e. the overall size, position, and orientation of objects in the images) will react to changes in volume and tempo. The higher the number, the higher the sensitivity.

Recommended range: 0.05 – 0.8

Default: 0.25

Example:

python visualize.py --song beethoven.mp3 --pitch_sensitivity 0.1

depth

The depth specifies the max value of the class vector units. Numbers closer to 1 seem to yield more thematically rich content. Numbers closer to 0 seem to yield more 'deep' structures like human and dog faces. However, this depends heavily on the specific classes you are using.

Range: 0.01 - 1

Default: 1

Example:

python visualize.py --song beethoven.mp3 --depth 0.5

classes

If you want to choose which classes (image categories) to visualize, you can specify a list of ImageNet indices (1-1000) here. (list of ImageNet class indices). The number of classes must be equal to [num_classes] (default is twelve, corresponding to the twelve musical pitches (A, A#, B, etc.)). You can also enter the class indices in order of priority (highest priority first) and set sort_classes_by_power to 1.

Default: Twelve random indices between 0-999

Example (if num_classes is set to default of twelve):

python visualize.py --song beethoven.mp3 --classes 45 99 567 234 89 90 105 998 56 677 884 530

num_classes

If you want to focus the visualizer around fewer than twelve themes, you can set num_classes to a number less than twelve. Since each class is associated with a pitch, the pitches that are retained when num_classes < 12 are those with the most overall power in the song.

Default: 12

Example:

python visualize.py --song beethoven.mp3 --num_classes 4 

Or if you want to choose the classes:

python visualize.py --song beethoven.mp3 --num_classes 4 --classes 987 23 56 782

sort_classes_by_power

Set this to 1 if you want to prioritize the classes based on the order that you entered them in the class input. If you do not specify the class input, there is no reason to set this to 1. If you do specify the class input and do not set this to 1, the classes will be associated with the pitches in harmonic order from A, A#, B, etc.

Example:

python visualize.py --song beethoven.mp3  --classes 45 99 567 234 89 90 105 998 56 677 884 530 --sort_classes_by_power 1

jitter

The jitter prevents the same exact noise vectors from cycling repetitively during repetitive music so that the video output is more interesting. If you do want to cycle repetitively, set jitter to 0.

Range: 0 – 1 Default: 0.5

Example:

python visualize.py --song beethoven.mp3 --jitter 0

frame_length

The frame length controls the number of audio frames per video frame in the output. If you want a higher frame rate for visualizing very rapid music, lower the frame_length. If you want a lower frame rate (perhaps if you are running on a CPU and want to cut down your runtime), raise the frame_length. The default of 512 is high quality.

Range: Multiples of 2^6

Default: 512

Example:

python visualize.py --song beethoven.mp3 --frame_length 2048

truncation

The truncation controls the variability of images that BigGAN generates by limiting the max values in the noise vector. Truncations closer to 1 yield more variable images, and truncations closer to 0 yield simpler images with more recognizable, normal-looking objects.

Range: 0.1 - 1

Default: 1

Example:

python visualize.py --song beethoven.mp3 --truncation 0.4

smooth_factor

After the class vectors have been generated, they are smoothed by interpolating linearly between the means of class vectors in bins of size [smooth_factor]. This is performed because small local fluctuations in pitch can cause the video frames to fluctuate back and forth. If you want to visualize very fast music with rapid changes in pitch, you can lower the smooth factor. You may also want to lower the frame_length in that case. However, for most songs, it is difficult to avoid rapid fluctuations with smooth factors less than 10.

Range: > 0

Recommended range: 10 – 30

Default: 20

Example:

python visualize.py --song beethoven.mp3 --smooth_factor 6

batch_size

BigGAN generates the images in batches of size [batch_size]. The only reason to reduce batch size from the default of 30 is if you run out of CUDA memory on a GPU. Reducing the batch size will slightly increase overall runtime.

Default: 30

Example:

python visualize.py --song beethoven.mp3 --batch_size 20

use_previous_classes

If your previous run of the visualizer used random classes (i.e. you did not manually set the class input), and you liked the video output but want to mess with some other parameters, set use_previous_classes to 1 so that you create a similar video with the same classes on the next run of the code.

Default: 0

Example:

python visualize.py --song beethoven.mp3 --use_previous_classes 1

use_previous_vectors

If you're messing around with the visualizer parameters, it can be useful to generate videos with short durations to keep your runtime low. Once you find the right set of parameters, remove the duration argument and set use_previous_vectors to 1 to generate the same video again, but for a longer duration.

Default: 0

Example:

python visualize.py --song beethoven.mp3 --use_previous_vectors 1

output_file

Output file name

Default: output.mp4

Example:

python visualize.py --song beethoven.mp3 --output_file my_movie.mp4

deep-music-visualizer's People

Contributors

msieg avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.