diff --git a/MANIFEST.in b/MANIFEST.in index 861cb5a794..0d8b4b4ce4 100644 --- a/MANIFEST.in +++ b/MANIFEST.in @@ -1,6 +1,7 @@ include README.md include LICENSE.txt include requirements.*.txt +include requirements.txt include TTS/VERSION recursive-include TTS *.json recursive-include TTS *.html diff --git a/Makefile b/Makefile index c7815f1917..d3d7dd416b 100644 --- a/Makefile +++ b/Makefile @@ -34,6 +34,9 @@ dev-deps: ## install development deps doc-deps: ## install docs dependencies pip install -r docs/requirements.txt +build-docs: ## build the docs + cd docs && make clean && make build + hub-deps: ## install deps for torch hub use pip install -r requirements.hub.txt diff --git a/TTS/.models.json b/TTS/.models.json index b8d8d4f711..73204db64f 100644 --- a/TTS/.models.json +++ b/TTS/.models.json @@ -51,7 +51,7 @@ "vctk":{ "sc-glow-tts": { "description": "Multi-Speaker Transformers based SC-Glow model from https://arxiv.org/abs/2104.05557.", - "github_rls_url": "https://github.com/coqui-ai/TTS/releases/download/v0.1.0/tts_models--en--vctk--sc-glowtts-transformer.zip", + "github_rls_url": "https://github.com/coqui-ai/TTS/releases/download/v0.1.0/tts_models--en--vctk--sc-glow-tts.zip", "default_vocoder": "vocoder_models/en/vctk/hifigan_v2", "commit": "b531fa69", "author": "Edresson Casanova", diff --git a/TTS/VERSION b/TTS/VERSION index 6c6aa7cb09..6da28dde76 100644 --- a/TTS/VERSION +++ b/TTS/VERSION @@ -1 +1 @@ -0.1.0 \ No newline at end of file +0.1.1 \ No newline at end of file diff --git a/TTS/utils/trainer_utils.py b/TTS/utils/trainer_utils.py index 02e68905e3..29915527fb 100644 --- a/TTS/utils/trainer_utils.py +++ b/TTS/utils/trainer_utils.py @@ -11,11 +11,15 @@ def is_apex_available(): def setup_torch_training_env(cudnn_enable, cudnn_benchmark): + num_gpus = torch.cuda.device_count() + if num_gpus > 1: + raise RuntimeError( + f" [!] {num_gpus} active GPUs. Define the target GPU by `CUDA_VISIBLE_DEVICES`. For multi-gpu training use `TTS/bin/distribute.py`." + ) torch.backends.cudnn.enabled = cudnn_enable torch.backends.cudnn.benchmark = cudnn_benchmark torch.manual_seed(54321) use_cuda = torch.cuda.is_available() - num_gpus = torch.cuda.device_count() print(" > Using CUDA: ", use_cuda) print(" > Number of GPUs: ", num_gpus) return use_cuda, num_gpus diff --git a/tests/inference_tests/test_synthesize.py b/tests/inference_tests/test_synthesize.py index 62eb6dbe61..526f7dc809 100644 --- a/tests/inference_tests/test_synthesize.py +++ b/tests/inference_tests/test_synthesize.py @@ -10,19 +10,19 @@ def test_synthesize(): # single speaker model run_cli(f'tts --text "This is an example." --out_path "{output_path}"') - # run_cli( - # "tts --model_name tts_models/en/ljspeech/speedy-speech-wn " - # f'--text "This is an example." --out_path "{output_path}"' - # ) - # run_cli( - # "tts --model_name tts_models/en/ljspeech/speedy-speech-wn " - # "--vocoder_name vocoder_models/en/ljspeech/multiband-melgan " - # f'--text "This is an example." --out_path "{output_path}"' - # ) + run_cli( + "tts --model_name tts_models/en/ljspeech/speedy-speech-wn " + f'--text "This is an example." --out_path "{output_path}"' + ) + run_cli( + "tts --model_name tts_models/en/ljspeech/speedy-speech-wn " + "--vocoder_name vocoder_models/en/ljspeech/multiband-melgan " + f'--text "This is an example." --out_path "{output_path}"' + ) - # # multi-speaker model - # run_cli("tts --model_name tts_models/en/vctk/sc-glow-tts --list_speaker_idxs") - # run_cli( - # f'tts --model_name tts_models/en/vctk/sc-glow-tts --speaker_idx "p304" ' - # f'--text "This is an example." --out_path "{output_path}"' - # ) + # multi-speaker model + run_cli("tts --model_name tts_models/en/vctk/sc-glow-tts --list_speaker_idxs") + run_cli( + f'tts --model_name tts_models/en/vctk/sc-glow-tts --speaker_idx "p304" ' + f'--text "This is an example." --out_path "{output_path}"' + )