Skip to content

2.0:Greatly optimized training speed

Compare
Choose a tag to compare
@yxlllc yxlllc released this 21 Mar 16:46
· 255 commits to master since this release

Unzip the pretrained model into exp directory, then run the demo samples:

# opencpop (1st speaker)
python main.py -i samples/source.wav -m exp/multi_speaker/model_300000.pt -o samples/svc-opencpop+12key.wav -k 12 -id 1
# kiritan (2nd speaker)
python main.py -i samples/source.wav -m exp/multi_speaker/model_300000.pt -o samples/svc-kiritan+12key.wav -k 12 -id 2
# mix the timbre of opencpop and kiritan in a 0.5 to 0.5 ratio
python main.py -i samples/source.wav -m exp/multi_speaker/model_300000.pt -o samples/svc-opencpop_kiritan_mix+12key.wav -k 12 -mix "{1:0.5, 2:0.5}"

The training data of this 2-speaker model is from opencpop and kiritan

Thanks to CN_ChiTu for helping to train this model.