Spaces:
Running
Running
import sys | |
import os | |
import torch | |
# By using XTTS you agree to CPML license https://coqui.ai/cpml | |
os.environ["COQUI_TOS_AGREED"] = "1" | |
import gradio as gr | |
from TTS.api import TTS | |
from TTS.utils.manage import ModelManager | |
model_names = TTS().list_models() | |
print(model_names.__dict__) | |
print(model_names.__dir__()) | |
model_name = "tts_models/multilingual/multi-dataset/xtts_v2" | |
m = model_name | |
# Automatic device detection | |
if torch.cuda.is_available(): | |
# cuda only | |
device_type = "cuda" | |
device_selection = "cuda:0" | |
data_type = torch.float16 | |
else: | |
# no GPU or Amd | |
device_type = "cpu" | |
device_selection = "cpu" | |
data_type = torch.float32 | |
tts = TTS(model_name, gpu=torch.cuda.is_available()) | |
tts.to(device_type) | |
def predict(prompt, language, audio_file_pth, mic_file_path, use_mic): | |
if True: | |
if use_mic: | |
if mic_file_path is None: | |
gr.Warning("Please record your voice with Microphone, or uncheck Use Microphone to use reference audios") | |
return ( | |
None, | |
None, | |
) | |
else: | |
speaker_wav = mic_file_path | |
else: | |
speaker_wav = audio_file_pth | |
if len(prompt) < 2: | |
gr.Warning("Please give a longer prompt text") | |
return ( | |
None, | |
None, | |
) | |
if len(prompt) > 50000: | |
gr.Warning("Text length limited to 50000 characters for this demo, please try shorter text") | |
return ( | |
None, | |
None, | |
) | |
try: | |
if language == "fr": | |
if m.find("your") != -1: | |
language = "fr-fr" | |
if m.find("/fr/") != -1: | |
language = None | |
tts.tts_to_file( | |
text=prompt, | |
file_path="output.wav", | |
speaker_wav=speaker_wav, | |
language=language | |
) | |
except RuntimeError as e : | |
if "device-assert" in str(e): | |
# cannot do anything on cuda device side error, need to restart | |
gr.Warning("Unhandled Exception encounter, please retry in a minute") | |
print("Cuda device-assert Runtime encountered need restart") | |
sys.exit("Exit due to cuda device-assert") | |
else: | |
raise e | |
return ( | |
gr.make_waveform( | |
audio="output.wav", | |
), | |
"output.wav", | |
) | |
title = "Multi-language Text-to-Speech" | |
description = f""" | |
<a href="https://huggingface.co/coqui/XTTS-v1">XTTS</a> is a Voice generation model that lets you clone voices into different languages by using just a quick 3-second audio clip. | |
<br/> | |
XTTS is built on previous research, like Tortoise, with additional architectural innovations and training to make cross-language voice cloning and multilingual speech generation possible. | |
<br/> | |
This is the same model that powers our creator application <a href="https://coqui.ai">Coqui Studio</a> as well as the <a href="https://docs.coqui.ai">Coqui API</a>. In production we apply modifications to make low-latency streaming possible. | |
<br/> | |
Leave a star on the Github <a href="https://github.com/coqui-ai/TTS">TTS</a>, where our open-source inference and training code lives. | |
<br/> | |
<p>For faster inference without waiting in the queue, you should duplicate this space and upgrade to GPU via the settings. | |
<br/> | |
<a href="https://huggingface.co/spaces/coqui/xtts?duplicate=true"> | |
<img style="margin-top: 0em; margin-bottom: 0em" src="https://bit.ly/3gLdBN6" alt="Duplicate Space"></a> | |
</p> | |
""" | |
article = "" | |
examples = [ | |
] | |
gr.Interface( | |
fn=predict, | |
inputs=[ | |
gr.Textbox( | |
label="Text Prompt", | |
info="One or two sentences at a time is better", | |
value="Hello, World! Here is an example of light voice cloning. Try to upload your best audio samples quality", | |
), | |
gr.Dropdown( | |
label="Language", | |
info="Select an output language for the synthesised speech", | |
choices=[ | |
["Arabic", "ar"], | |
["Brazilian Portuguese", "pt"], | |
["Mandarin Chinese", "zh-cn"], | |
["Czech", "cs"], | |
["Dutch", "nl"], | |
["English", "en"], | |
["French", "fr"], | |
["German", "de"], | |
["Italian", "it"], | |
["Polish", "pl"], | |
["Russian", "ru"], | |
["Spanish", "es"], | |
["Turkish", "tr"] | |
], | |
max_choices=1, | |
value="en", | |
), | |
gr.Audio( | |
label="Reference Audio", | |
#info="Click on the ✎ button to upload your own target speaker audio", | |
type="filepath", | |
value="examples/female.wav", | |
), | |
gr.Audio(sources=["microphone"], | |
type="filepath", | |
#info="Use your microphone to record audio", | |
label="Use Microphone for Reference"), | |
gr.Checkbox(label="Check to use Microphone as Reference", | |
value=False, | |
info="Notice: Microphone input may not work properly under traffic",), | |
], | |
outputs=[ | |
gr.Video(label="Waveform Visual", autoplay=True), | |
gr.Audio(label="Synthesised Audio", autoplay=False), | |
], | |
title=title, | |
description=description, | |
article=article, | |
examples=examples, | |
).queue().launch(debug=True) |