Skip to content

Commit 55e2f46

Browse files
1. add huggingface SVS; 2. add inference logic from raw inputs; 3. update readme.
1 parent 368719e commit 55e2f46

2 files changed

Lines changed: 3 additions & 3 deletions

File tree

README.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -21,10 +21,10 @@ This repository is the official PyTorch implementation of our AAAI-2022 [paper](
2121

2222
:tada: :tada: :tada: **Updates**:
2323
- Jul.27, 2022: Update documents for [SVS](docs/README-SVS.md). Add [inference logic from raw input](docs/README-SVS-opencpop-cascade.md#4-inference-from-raw-inputs) for SVS; Add Interactive SVS running on [HuggingFace🤗 SVS](https://huggingface.co/spaces/Silentlin/DiffSinger).
24-
- Mar.2, 2022: MIDI-new-version.
24+
- Mar.2, 2022: MIDI-B-version.
2525
- Mar.1, 2022: [NeuralSVB](https://github.com/MoonInTheRiver/NeuralSVB), for singing voice beautifying, has been released.
2626
- Feb.13, 2022: [NATSpeech](https://github.com/NATSpeech/NATSpeech), the improved code framework, which contains the implementations of DiffSpeech and our NeurIPS-2021 work [PortaSpeech](https://openreview.net/forum?id=xmJsuh8xlq) has been released.
27-
- Jan.29, 2022: support MIDI-old-version SVS.
27+
- Jan.29, 2022: support MIDI-A-version SVS.
2828
- Jan.13, 2022: support SVS, release PopCS dataset.
2929
- Dec.19, 2021: support TTS. [HuggingFace🤗 TTS](https://huggingface.co/spaces/NATSpeech/DiffSpeech)
3030

docs/README-SVS-popcs.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -54,7 +54,7 @@ Remember to put the pre-trained models in `checkpoints` directory.
5454
*Note that:*
5555

5656
- *the original PWG version vocoder in the paper we used has been put into commercial use, so we provide this HifiGAN version vocoder as a substitute.*
57-
- *we assume the ground-truth F0 to be given as the pitch information following [1][2][3]. If you want to conduct experiments on MIDI data, you need an external F0 predictor (like [MIDI-old-version](README-SVS-opencpop-cascade.md)) or a joint prediction with spectrograms(like [MIDI-new-version](README-SVS-opencpop-e2e.md)).*
57+
- *we assume the ground-truth F0 to be given as the pitch information following [1][2][3]. If you want to conduct experiments on MIDI data, you need an external F0 predictor (like [MIDI-A-version](README-SVS-opencpop-cascade.md)) or a joint prediction with spectrograms(like [MIDI-B-version](README-SVS-opencpop-e2e.md)).*
5858

5959
[1] Adversarially trained multi-singer sequence-to-sequence singing synthesizer. Interspeech 2020.
6060

0 commit comments

Comments
 (0)