/amt-apc

音乐: 自动钢琴翻唱: AMT-APC: AMT-APC: Automatic Piano Cover by Fine-Tuning an Automatic Music Transcription Model

Primary LanguagePythonMIT LicenseMIT

AMT-APC

AMT-APC is a method for training an automatic piano cover generation model by fine-tuning an AMT (Automatic Music Transcription) model.

Usage (Piano Cover Generation)

Python version: 3.10

  1. Install dependencies
pip install -r requirements.txt

Alternatively, if you only need to run the inference code, you can install just the necessary packages.

pip install torch torchaudio pretty-midi tqdm
  1. Download the pre-trained model
wget -P models/params/ https://github.com/misya11p/amt-apc/releases/download/beta/apc.pth
  1. Run the inference code
python infer input.wav

You can also input a YouTube URL (requires yt-dlp).

python infer 'https://www.youtube.com/watch?v=...'

You can also specify a style (level1, level2, level3).

python infer input.wav --style level3

Usage (Training & Evaluation)

Python version: 3.10

  1. Install dependencies
pip install -r requirements.txt
  1. Download the pre-trained AMT model
wget -P models/params/ https://github.com/misya11p/amt-apc/releases/download/beta/amt.pth
  1. Download the dataset
python download.py

The dataset directory is set to dataset/ by default. You can change this directory by modifying path.dataset in config.json.

  1. Create the dataset
python data/sync.py
python data/transcribe.py
python data/sv/extract.py
python data/create_labels.py
python data/create_dataset.py
  1. Train the model
python train --n_gpus 1
  1. Evaluate the model

Calculate $\mathcal Q_{\text{max}}$.

git clone https://github.com/albincorreya/ChromaCoverId.git eval/ChromaCoverId
python eval/cover.py
python eval/distance.py

Options

Detailed configuration can be done through config.json or by using command line options, which are explained with --help. The default values are those used in the experiments in the paper.