Update app.py
Browse files
app.py
CHANGED
@@ -41,6 +41,10 @@ frame_asr = FrameBatchMultiTaskAED(
|
|
41 |
|
42 |
amp_dtype = torch.float16
|
43 |
|
|
|
|
|
|
|
|
|
44 |
def convert_audio(audio_filepath, tmpdir, utt_id):
|
45 |
"""
|
46 |
Convert all files to monochannel 16 kHz wav files.
|
@@ -220,6 +224,12 @@ def on_src_or_tgt_lang_change(src_lang_value, tgt_lang_value, pnc_value):
|
|
220 |
)
|
221 |
return src_lang, tgt_lang, pnc
|
222 |
|
|
|
|
|
|
|
|
|
|
|
|
|
223 |
|
224 |
with gr.Blocks(
|
225 |
title="MyAlexa",
|
@@ -282,7 +292,7 @@ with gr.Blocks(
|
|
282 |
|
283 |
|
284 |
go_button.click(
|
285 |
-
fn=
|
286 |
inputs = [audio_file, src_lang, tgt_lang, pnc],
|
287 |
outputs = [model_output_text_box]
|
288 |
)
|
|
|
41 |
|
42 |
amp_dtype = torch.float16
|
43 |
|
44 |
+
llm_pipeline = transformers.pipeline(
|
45 |
+
"text-generation", model="meta-llama/Meta-Llama-3-8B", model_kwargs={"torch_dtype": torch.bfloat16}, device_map="auto"
|
46 |
+
)
|
47 |
+
|
48 |
def convert_audio(audio_filepath, tmpdir, utt_id):
|
49 |
"""
|
50 |
Convert all files to monochannel 16 kHz wav files.
|
|
|
224 |
)
|
225 |
return src_lang, tgt_lang, pnc
|
226 |
|
227 |
+
def main(audio_filepath, src_lang, tgt_lang, pnc):
|
228 |
+
translated = transcribe(audio_filepath, src_lang, tgt_lang, pnc)
|
229 |
+
answer = llm_pipeline(translated)
|
230 |
+
return answer
|
231 |
+
|
232 |
+
|
233 |
|
234 |
with gr.Blocks(
|
235 |
title="MyAlexa",
|
|
|
292 |
|
293 |
|
294 |
go_button.click(
|
295 |
+
fn=main,
|
296 |
inputs = [audio_file, src_lang, tgt_lang, pnc],
|
297 |
outputs = [model_output_text_box]
|
298 |
)
|