This repository demonstrates how to run NVIDIA's Flowtron text-to-speech synthesis model on an RTX 4090 GPU using Docker and Docker Compose. The setup leverages GPU acceleration to efficiently handle the computational load of training and inference. Docker image has all model checkpoints and sample data downloaded to allow for straight forward inference_style_transfer.ipynb
notebook execution.
- NVIDIA GPU: Ensure you have an NVIDIA GPU (e.g., RTX 4090) with the latest drivers installed.
- Docker: Install Docker on your machine. Instructions can be found here.
- Docker Compose: Install Docker Compose. Instructions can be found here.
-
Clone the Repository:
git clone https://github.com/aleksas/flowtron.git cd flowtron
-
Build the Docker Image: Use the provided
Dockerfile
anddocker-compose.yml
to build the Docker image.docker-compose build
-
Run the Container: Start the container with GPU support.
docker-compose up
-
Access the Jupyter Notebook: Open your web browser and navigate to
http://localhost:8898
to access the Jupyter notebook interface. You can start experimenting with Flowtron using the GPU-accelerated environment.
Dockerfile
: Defines the Docker image, including all dependencies and necessary setup.docker-compose.yml
: Orchestrates the container setup, ensuring the correct ports and GPU settings.- Pre-trained models: Automatically downloaded during the build process.
- Original Flowtron repository by NVIDIA: https://github.com/NVIDIA/flowtron
- This setup is configured and tested to work on RTX 4090, enabling faster processing and experimentation.
This project is licensed under the MIT License. See the original Flowtron repository for more details.