πViTTS is a library for advanced Text-to-Speech generation for multi language such as: chinese, japanese, vietnamese .
πViTTS builts on the latest research, It designed to achieve the best trade-off among ease of training, inference and evaluate.
πViTTS is a library for text to speech, it achive performance in speech and quality.
Please use our dedicated channels for questions and discussion. Help is much more valuable if it's shared publicly so that more people can benefit from it.
Type | Platforms |
---|---|
π¨ Bug Reports | GitHub Issue |
π Feature Requests & Ideas | GitHub Issue |
π©βπ» Usage Questions | Github Discussions |
π― General Discussion | Linkedin or Gitter Room |
Type | Links |
---|---|
πΌ Documentation | ReadTheDocs |
πΎ Installation | TTS/README.md |
π©βπ» Contributing | CONTRIBUTING.md |
π Road Map | Main Development Plans |
- Tacotron: paper
- Tacotron2: paper
- Glow-TTS: paper
- Speedy-Speech: paper
- Align-TTS: paper
- FastPitch: paper
- FastSpeech: paper
- VITS: paper
- Guided Attention: paper
- Forward Backward Decoding: paper
- Graves Attention: paper
- Double Decoder Consistency: blog
- Dynamic Convolutional Attention: paper
- Alignment Network: paper
- MelGAN: paper
- MultiBandMelGAN: paper
- ParallelWaveGAN: paper
- GAN-TTS discriminators: paper
- WaveRNN: origin
- WaveGrad: paper
- HiFiGAN: paper
- UnivNet: paper
You can also help us implement more models.