mrfakename commited on
Commit
083a2d9
·
verified ·
1 Parent(s): fbe6497

Sync from GitHub repo

Browse files

This Space is synced from the GitHub repo: https://github.com/SWivid/F5-TTS. Please submit contributions to the Space there

README_REPO.md CHANGED
@@ -114,9 +114,11 @@ Deployment solution with Triton and TensorRT-LLM.
114
  #### Benchmark Results
115
  Decoding on a single L20 GPU, using 26 different prompt_audio & target_text pairs.
116
 
117
- | Model | Concurrency | Avg Latency | RTF |
118
- |-------|-------------|----------------|-------|
119
- | F5-TTS Base (Vocos) | 1 | 253 ms | 0.0394|
 
 
120
 
121
  See [detailed instructions](src/f5_tts/runtime/triton_trtllm/README.md) for more information.
122
 
 
114
  #### Benchmark Results
115
  Decoding on a single L20 GPU, using 26 different prompt_audio & target_text pairs.
116
 
117
+ | Model | Concurrency | Avg Latency | RTF | Mode |
118
+ |---------------------|----------------|-------------|--------|-----------------|
119
+ | F5-TTS Base (Vocos) | 2 | 253 ms | 0.0394 | Client-Server |
120
+ | F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.0402 | Offline TRT-LLM |
121
+ | F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.1467 | Offline Pytorch |
122
 
123
  See [detailed instructions](src/f5_tts/runtime/triton_trtllm/README.md) for more information.
124
 
pyproject.toml CHANGED
@@ -4,7 +4,7 @@ build-backend = "setuptools.build_meta"
4
 
5
  [project]
6
  name = "f5-tts"
7
- version = "1.1.0"
8
  description = "F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching"
9
  readme = "README.md"
10
  license = {text = "MIT License"}
 
4
 
5
  [project]
6
  name = "f5-tts"
7
+ version = "1.1.1"
8
  description = "F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching"
9
  readme = "README.md"
10
  license = {text = "MIT License"}
src/f5_tts/runtime/triton_trtllm/README.md CHANGED
@@ -59,11 +59,11 @@ benchmark.py --output-dir $log_dir \
59
  ### Benchmark Results
60
  Decoding on a single L20 GPU, using 26 different prompt_audio/target_text pairs.
61
 
62
- | Model | Concurrency | Avg Latency | RTF | Mode |
63
- |-------|-------------|----------------|-------|------|
64
- | F5-TTS Base (Vocos) | 2 | 253 ms | 0.0394|Client-Server|
65
- | F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.0402|Offline TRT-LLM|
66
- | F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.1467|Offline Pytorch|
67
 
68
  ### Credits
69
  1. [F5-TTS-TRTLLM](https://github.com/Bigfishering/f5-tts-trtllm)
 
59
  ### Benchmark Results
60
  Decoding on a single L20 GPU, using 26 different prompt_audio/target_text pairs.
61
 
62
+ | Model | Concurrency | Avg Latency | RTF | Mode |
63
+ |---------------------|----------------|-------------|--------|-----------------|
64
+ | F5-TTS Base (Vocos) | 2 | 253 ms | 0.0394 | Client-Server |
65
+ | F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.0402 | Offline TRT-LLM |
66
+ | F5-TTS Base (Vocos) | 1 (Batch_size) | - | 0.1467 | Offline Pytorch |
67
 
68
  ### Credits
69
  1. [F5-TTS-TRTLLM](https://github.com/Bigfishering/f5-tts-trtllm)