| --- |
| license: mit |
| --- |
| |
| # Amphion Vocoder Pretrained Models |
|
|
| We provide a [DiffWave](https://github.com/open-mmlab/Amphion/tree/main/egs/vocoder/diffusion) pretrained checkpoint, which is trained on 125 hours of speech data and 80 hours of singing voice data. |
|
|
| ## Quick Start |
|
|
| To utilize these pretrained vocoders, just run the following commands: |
|
|
| ### Step1: Download the checkpoint |
| ```bash |
| git lfs install |
| git clone https://huggingface.co/amphion/diffwave |
| ``` |
|
|
| ### Step2: Clone the Amphion's Source Code of GitHub |
| ```bash |
| git clone https://github.com/open-mmlab/Amphion.git |
| ``` |
|
|
| ### Step3: Specify the checkpoint's path |
| Use the soft link to specify the downloaded checkpoint in the first step: |
|
|
| ```bash |
| cd Amphion |
| mkdir -p ckpts/vocoder |
| ln -s "$(realpath ../diffwave/diffwave)" pretrained/diffwave |
| ``` |
|
|
| ### Step4: Inference |
| For analysis synthesis on the processed dataset, raw waveform, or predicted mel spectrograms, you can follow the inference part of [this recipe](https://github.com/open-mmlab/Amphion/tree/main/egs/vocoder/diffusion). |
|
|
| ```bash |
| sh egs/vocoder/diffusion/diffwave/run.sh --stage 3 \ |
| --infer_mode [Your chosen inference mode] \ |
| --infer_datasets [Datasets you want to inference, needed when infer_from_dataset] \ |
| --infer_feature_dir [Your path to your predicted acoustic features, needed when infer_from_feature] \ |
| --infer_audio_dir [Your path to your audio files, needed when infer_form_audio] \ |
| --infer_expt_dir Amphion/ckpts/vocoder/[YourExptName] \ |
| --infer_output_dir Amphion/ckpts/vocoder/[YourExptName]/result \ |
| ``` |