Update app.py
Browse files
app.py
CHANGED
@@ -19,7 +19,7 @@ def forward_gpu(ps, ref_s, speed):
|
|
19 |
return models[True](ps, ref_s, speed)
|
20 |
|
21 |
def generate_first(text, voice='af_heart', speed=1, use_gpu=CUDA_AVAILABLE):
|
22 |
-
text = text
|
23 |
pipeline = pipelines[voice[0]]
|
24 |
pack = pipeline.load_voice(voice)
|
25 |
use_gpu = use_gpu and CUDA_AVAILABLE
|
@@ -50,7 +50,7 @@ def tokenize_first(text, voice='af_heart'):
|
|
50 |
return words # Return a list of words
|
51 |
|
52 |
def generate_all(text, voice='af_heart', speed=1, use_gpu=CUDA_AVAILABLE):
|
53 |
-
text = text
|
54 |
pipeline = pipelines[voice[0]]
|
55 |
pack = pipeline.load_voice(voice)
|
56 |
use_gpu = use_gpu and CUDA_AVAILABLE
|
|
|
19 |
return models[True](ps, ref_s, speed)
|
20 |
|
21 |
def generate_first(text, voice='af_heart', speed=1, use_gpu=CUDA_AVAILABLE):
|
22 |
+
text = text
|
23 |
pipeline = pipelines[voice[0]]
|
24 |
pack = pipeline.load_voice(voice)
|
25 |
use_gpu = use_gpu and CUDA_AVAILABLE
|
|
|
50 |
return words # Return a list of words
|
51 |
|
52 |
def generate_all(text, voice='af_heart', speed=1, use_gpu=CUDA_AVAILABLE):
|
53 |
+
text = text
|
54 |
pipeline = pipelines[voice[0]]
|
55 |
pack = pipeline.load_voice(voice)
|
56 |
use_gpu = use_gpu and CUDA_AVAILABLE
|