Stardust-minus commited on
Commit
9c347e6
·
verified ·
1 Parent(s): f65fe2e

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -41,6 +41,3 @@ examples/English.wav filter=lfs diff=lfs merge=lfs -text
41
  examples/French.wav filter=lfs diff=lfs merge=lfs -text
42
  examples/German.wav filter=lfs diff=lfs merge=lfs -text
43
  examples/Spanish.wav filter=lfs diff=lfs merge=lfs -text
44
- 022b2161-8f56-4432-a9ae-b4bd514e4821.mp3 filter=lfs diff=lfs merge=lfs -text
45
- output.wav filter=lfs diff=lfs merge=lfs -text
46
- ref.wav filter=lfs diff=lfs merge=lfs -text
 
41
  examples/French.wav filter=lfs diff=lfs merge=lfs -text
42
  examples/German.wav filter=lfs diff=lfs merge=lfs -text
43
  examples/Spanish.wav filter=lfs diff=lfs merge=lfs -text
 
 
 
app.py CHANGED
@@ -313,4 +313,4 @@ if __name__ == "__main__":
313
  inference_fct = get_inference_wrapper(inference_engine)
314
 
315
  app = build_app(inference_fct, args.theme)
316
- app.queue(api_open=True).launch(show_error=True, show_api=True)
 
313
  inference_fct = get_inference_wrapper(inference_engine)
314
 
315
  app = build_app(inference_fct, args.theme)
316
+ app.queue(api_open=True).launch(show_error=True, show_api=True, server_name="0.0.0.0", server_port=18888)
fish_speech/models/text2semantic/llama.py CHANGED
@@ -48,7 +48,7 @@ class BaseModelArgs:
48
 
49
  # Codebook configs
50
  codebook_size: int = 160
51
- num_codebooks: int = 9
52
 
53
  # Gradient checkpointing
54
  use_gradient_checkpointing: bool = True
 
48
 
49
  # Codebook configs
50
  codebook_size: int = 160
51
+ num_codebooks: int = 4
52
 
53
  # Gradient checkpointing
54
  use_gradient_checkpointing: bool = True
tools/api.py CHANGED
@@ -136,7 +136,7 @@ async def other_exception_handler(exc: "Exception"):
136
 
137
 
138
  def load_audio(reference_audio, sr):
139
- if len(str(reference_audio)) > 255 or not Path(reference_audio).exists():
140
  audio_data = reference_audio
141
  reference_audio = io.BytesIO(audio_data)
142
 
 
136
 
137
 
138
  def load_audio(reference_audio, sr):
139
+ if len(reference_audio) > 255 or not Path(reference_audio).exists():
140
  audio_data = reference_audio
141
  reference_audio = io.BytesIO(audio_data)
142
 
tools/vqgan/inference.py CHANGED
@@ -14,7 +14,7 @@ from omegaconf import OmegaConf
14
  from tools.file import AUDIO_EXTENSIONS
15
 
16
  # register eval resolver
17
- #OmegaConf.register_new_resolver("eval", eval)
18
 
19
 
20
  def load_model(config_name, checkpoint_path, device="cuda"):
 
14
  from tools.file import AUDIO_EXTENSIONS
15
 
16
  # register eval resolver
17
+ OmegaConf.register_new_resolver("eval", eval)
18
 
19
 
20
  def load_model(config_name, checkpoint_path, device="cuda"):