Spaces:
Running
on
L4
Running
on
L4
Upload folder using huggingface_hub
Browse files- .gitattributes +0 -3
- app.py +1 -1
- fish_speech/models/text2semantic/llama.py +1 -1
- tools/api.py +1 -1
- tools/vqgan/inference.py +1 -1
.gitattributes
CHANGED
@@ -41,6 +41,3 @@ examples/English.wav filter=lfs diff=lfs merge=lfs -text
|
|
41 |
examples/French.wav filter=lfs diff=lfs merge=lfs -text
|
42 |
examples/German.wav filter=lfs diff=lfs merge=lfs -text
|
43 |
examples/Spanish.wav filter=lfs diff=lfs merge=lfs -text
|
44 |
-
022b2161-8f56-4432-a9ae-b4bd514e4821.mp3 filter=lfs diff=lfs merge=lfs -text
|
45 |
-
output.wav filter=lfs diff=lfs merge=lfs -text
|
46 |
-
ref.wav filter=lfs diff=lfs merge=lfs -text
|
|
|
41 |
examples/French.wav filter=lfs diff=lfs merge=lfs -text
|
42 |
examples/German.wav filter=lfs diff=lfs merge=lfs -text
|
43 |
examples/Spanish.wav filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
app.py
CHANGED
@@ -313,4 +313,4 @@ if __name__ == "__main__":
|
|
313 |
inference_fct = get_inference_wrapper(inference_engine)
|
314 |
|
315 |
app = build_app(inference_fct, args.theme)
|
316 |
-
app.queue(api_open=True).launch(show_error=True, show_api=True)
|
|
|
313 |
inference_fct = get_inference_wrapper(inference_engine)
|
314 |
|
315 |
app = build_app(inference_fct, args.theme)
|
316 |
+
app.queue(api_open=True).launch(show_error=True, show_api=True, server_name="0.0.0.0", server_port=18888)
|
fish_speech/models/text2semantic/llama.py
CHANGED
@@ -48,7 +48,7 @@ class BaseModelArgs:
|
|
48 |
|
49 |
# Codebook configs
|
50 |
codebook_size: int = 160
|
51 |
-
num_codebooks: int =
|
52 |
|
53 |
# Gradient checkpointing
|
54 |
use_gradient_checkpointing: bool = True
|
|
|
48 |
|
49 |
# Codebook configs
|
50 |
codebook_size: int = 160
|
51 |
+
num_codebooks: int = 4
|
52 |
|
53 |
# Gradient checkpointing
|
54 |
use_gradient_checkpointing: bool = True
|
tools/api.py
CHANGED
@@ -136,7 +136,7 @@ async def other_exception_handler(exc: "Exception"):
|
|
136 |
|
137 |
|
138 |
def load_audio(reference_audio, sr):
|
139 |
-
if len(
|
140 |
audio_data = reference_audio
|
141 |
reference_audio = io.BytesIO(audio_data)
|
142 |
|
|
|
136 |
|
137 |
|
138 |
def load_audio(reference_audio, sr):
|
139 |
+
if len(reference_audio) > 255 or not Path(reference_audio).exists():
|
140 |
audio_data = reference_audio
|
141 |
reference_audio = io.BytesIO(audio_data)
|
142 |
|
tools/vqgan/inference.py
CHANGED
@@ -14,7 +14,7 @@ from omegaconf import OmegaConf
|
|
14 |
from tools.file import AUDIO_EXTENSIONS
|
15 |
|
16 |
# register eval resolver
|
17 |
-
|
18 |
|
19 |
|
20 |
def load_model(config_name, checkpoint_path, device="cuda"):
|
|
|
14 |
from tools.file import AUDIO_EXTENSIONS
|
15 |
|
16 |
# register eval resolver
|
17 |
+
OmegaConf.register_new_resolver("eval", eval)
|
18 |
|
19 |
|
20 |
def load_model(config_name, checkpoint_path, device="cuda"):
|