Spaces:
Running
on
Zero
Running
on
Zero
Sync from GitHub repo
Browse filesThis Space is synced from the GitHub repo: https://github.com/SWivid/F5-TTS. Please submit contributions to the Space there
- README_REPO.md +5 -3
- pyproject.toml +1 -1
- src/f5_tts/runtime/triton_trtllm/README.md +5 -5
README_REPO.md
CHANGED
@@ -114,9 +114,11 @@ Deployment solution with Triton and TensorRT-LLM.
|
|
114 |
#### Benchmark Results
|
115 |
Decoding on a single L20 GPU, using 26 different prompt_audio & target_text pairs.
|
116 |
|
117 |
-
| Model
|
118 |
-
|
119 |
-
| F5-TTS Base (Vocos) |
|
|
|
|
|
120 |
|
121 |
See [detailed instructions](src/f5_tts/runtime/triton_trtllm/README.md) for more information.
|
122 |
|
|
|
114 |
#### Benchmark Results
|
115 |
Decoding on a single L20 GPU, using 26 different prompt_audio & target_text pairs.
|
116 |
|
117 |
+
| Model | Concurrency | Avg Latency | RTF | Mode |
|
118 |
+
|---------------------|----------------|-------------|--------|-----------------|
|
119 |
+
| F5-TTS Base (Vocos) | 2 | 253 ms | 0.0394 | Client-Server |
|
120 |
+
| F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.0402 | Offline TRT-LLM |
|
121 |
+
| F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.1467 | Offline Pytorch |
|
122 |
|
123 |
See [detailed instructions](src/f5_tts/runtime/triton_trtllm/README.md) for more information.
|
124 |
|
pyproject.toml
CHANGED
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
|
|
4 |
|
5 |
[project]
|
6 |
name = "f5-tts"
|
7 |
-
version = "1.1.
|
8 |
description = "F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching"
|
9 |
readme = "README.md"
|
10 |
license = {text = "MIT License"}
|
|
|
4 |
|
5 |
[project]
|
6 |
name = "f5-tts"
|
7 |
+
version = "1.1.1"
|
8 |
description = "F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching"
|
9 |
readme = "README.md"
|
10 |
license = {text = "MIT License"}
|
src/f5_tts/runtime/triton_trtllm/README.md
CHANGED
@@ -59,11 +59,11 @@ benchmark.py --output-dir $log_dir \
|
|
59 |
### Benchmark Results
|
60 |
Decoding on a single L20 GPU, using 26 different prompt_audio/target_text pairs.
|
61 |
|
62 |
-
| Model
|
63 |
-
|
64 |
-
| F5-TTS Base (Vocos) | 2
|
65 |
-
| F5-TTS Base (Vocos) | 1 (Batch_size)
|
66 |
-
| F5-TTS Base (Vocos) | 1 (Batch_size)
|
67 |
|
68 |
### Credits
|
69 |
1. [F5-TTS-TRTLLM](https://github.com/Bigfishering/f5-tts-trtllm)
|
|
|
59 |
### Benchmark Results
|
60 |
Decoding on a single L20 GPU, using 26 different prompt_audio/target_text pairs.
|
61 |
|
62 |
+
| Model | Concurrency | Avg Latency | RTF | Mode |
|
63 |
+
|---------------------|----------------|-------------|--------|-----------------|
|
64 |
+
| F5-TTS Base (Vocos) | 2 | 253 ms | 0.0394 | Client-Server |
|
65 |
+
| F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.0402 | Offline TRT-LLM |
|
66 |
+
| F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.1467 | Offline Pytorch |
|
67 |
|
68 |
### Credits
|
69 |
1. [F5-TTS-TRTLLM](https://github.com/Bigfishering/f5-tts-trtllm)
|