synthviz
is a library for creating visualizations of piano-playing from MIDI files. The videos look like this:
entertainer.mp4
Right now, synthviz just provides a Python API. Command-line API is hopefully coming soon!
You'll need to install a couple of tools that make rendering this video possible:
- ffmpeg (creates video from audio and image video frames) - on Ubuntu,
sudo apt-get install ffmpeg
- timidity (synthesizes piano audio from MIDI) - on Ubuntu,
sudo apt-get install timidity
Install this package via pypi
:
pip install synthviz
You can use synthviz through the Python API:
from synthviz import create_video
create_video('river.midi') # provide str path of MIDI file
The create_video
function provides a lot of of options:
def create_video(input_midi: str,
video_filename = "output.mp4",
image_width = 1280,
image_height = 720,
black_key_height = 2/3,
falling_note_color = [75, 105, 177], # default: darker blue
pressed_key_color = [197, 208, 231], # default: lighter blue
vertical_speed = 1/4,
fps = 20
)
input_midi
(str): path to MIDI filevideo_filename
(str): path to output video, synthviz will write the video hereimage_width
(int): width of output video in pximage_height
(int): height of output video in pxblack_key_height
(float): height of black keys as a percentage of piano heightfalling_note_color
(Tuple[int]): color of falling keys in video, list of three RGB integerspressed_key_color
(Tuple[int]): color of pressed-down keys in video, list of three RGB integersvertical_speed
(float): the speed of the falling keys, fraction measured as main-image-heights per secondfps
(int): frames-per-second of output video
With the help of the piano_transcription_inference
library, you can make a cool video directly from raw audio!
First, install that library via pip install piano_transcription_inference
. Then run the following code:
import librosa
import os
import pathlib
from piano_transcription_inference import PianoTranscription, sample_rate, load_audio
from synthviz import create_video
audio_input = 'my_audio.mp3'
midi_intermediate_filename = 'transcription.mid'
video_filename = 'output.mp4'
transcriptor = PianoTranscription(device='cuda', checkpoint_path='./model.pth')
audio, _ = librosa.core.load(str(audio_input), sr=sample_rate)
transcribed_dict = transcriptor.transcribe(audio, midi_intermediate_filename)
create_video(input_midi=midi_intermediate_filename, video_filename=video_filename)
The synthviz library was originall adapted from this blog post, "Making Synthesia-style videos in Ubuntu", written by David Barry. Thanks David!