synthviz is a library for creating visualizations of piano-playing from MIDI files. The videos look like this:
entertainer.mp4
Right now, synthviz just provides a Python API. Command-line API is hopefully coming soon!
You'll need to install a couple of tools that make rendering this video possible:
- ffmpeg (creates video from audio and image video frames) - on Ubuntu,
sudo apt-get install ffmpeg - timidity (synthesizes piano audio from MIDI) - on Ubuntu,
sudo apt-get install timidity
Install this package via pypi:
pip install synthvizYou can use synthviz through the Python API:
from synthviz import create_video create_video('river.midi') # provide str path of MIDI fileThe create_video function provides a lot of of options:
def create_video(input_midi: str, video_filename = "output.mp4", image_width= 1280, image_height = 720, black_key_height = 2/3, falling_note_color = [75, 105, 177], # default: darker blue pressed_key_color = [197, 208, 231], # default: lighter blue vertical_speed = 1/4, fps = 20 ) input_midi(str): path to MIDI filevideo_filename(str): path to output video, synthviz will write the video hereimage_width(int): width of output video in pximage_height(int): height of output video in pxblack_key_height(float): height of black keys as a percentage of piano heightfalling_note_color(Tuple[int]): color of falling keys in video, list of three RGB integerspressed_key_color(Tuple[int]): color of pressed-down keys in video, list of three RGB integersvertical_speed(float): the speed of the falling keys, fraction measured as main-image-heights per secondfps(int): frames-per-second of output video
With the help of the piano_transcription_inference library, you can make a cool video directly from raw audio!
First, install that library via pip install piano_transcription_inference. Then run the following code:
import librosa import os import pathlib from piano_transcription_inference import PianoTranscription, sample_rate, load_audio from synthviz import create_video audio_input = 'my_audio.mp3' midi_intermediate_filename = 'transcription.mid' video_filename = 'output.mp4' transcriptor = PianoTranscription(device='cuda', checkpoint_path='./model.pth') audio, _ = librosa.core.load(str(audio_input), sr=sample_rate) transcribed_dict = transcriptor.transcribe(audio, midi_intermediate_filename) create_video(input_midi=midi_intermediate_filename, video_filename=video_filename)The synthviz library was originall adapted from this blog post, "Making Synthesia-style videos in Ubuntu", written by David Barry. Thanks David!