Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -346,41 +346,6 @@ def generate_answer(message, choice, retrieval_mode):
|
|
346 |
|
347 |
|
348 |
|
349 |
-
# def bot(history, choice, tts_choice, retrieval_mode):
|
350 |
-
# if not history:
|
351 |
-
# return history
|
352 |
-
|
353 |
-
# response, addresses = generate_answer(history[-1][0], choice, retrieval_mode)
|
354 |
-
# history[-1][1] = ""
|
355 |
-
|
356 |
-
# # Detect if the response is from Yelp (i.e., HTML formatted response)
|
357 |
-
# if "<table>" in response:
|
358 |
-
# for chunk in response.splitlines():
|
359 |
-
# history[-1][1] += chunk + "\n"
|
360 |
-
# time.sleep(0.1) # Adjust the delay as needed
|
361 |
-
# yield history, None
|
362 |
-
# return
|
363 |
-
|
364 |
-
# with concurrent.futures.ThreadPoolExecutor() as executor:
|
365 |
-
# if tts_choice == "Alpha":
|
366 |
-
# audio_future = executor.submit(generate_audio_elevenlabs, response)
|
367 |
-
# elif tts_choice == "Beta":
|
368 |
-
# audio_future = executor.submit(generate_audio_parler_tts, response)
|
369 |
-
# elif tts_choice == "Gamma":
|
370 |
-
# audio_future = executor.submit(generate_audio_mars5, response)
|
371 |
-
|
372 |
-
# for character in response:
|
373 |
-
# history[-1][1] += character
|
374 |
-
# time.sleep(0.05)
|
375 |
-
# yield history, None
|
376 |
-
|
377 |
-
# audio_path = audio_future.result()
|
378 |
-
# yield history, audio_path
|
379 |
-
|
380 |
-
# history.append([response, None]) # Ensure the response is added in the correct format
|
381 |
-
|
382 |
-
|
383 |
-
|
384 |
def bot(history, choice, tts_choice, retrieval_mode):
|
385 |
if not history:
|
386 |
return history
|
@@ -414,10 +379,8 @@ def bot(history, choice, tts_choice, retrieval_mode):
|
|
414 |
|
415 |
history.append([response, None]) # Ensure the response is added in the correct format
|
416 |
|
417 |
-
|
418 |
-
|
419 |
-
map_html = update_map_with_response(history)
|
420 |
-
yield history, map_html
|
421 |
|
422 |
|
423 |
|
@@ -564,26 +527,6 @@ base_audio_drive = "/data/audio"
|
|
564 |
|
565 |
#Normal Code with sample rate is 44100 Hz
|
566 |
|
567 |
-
# def transcribe_function(stream, new_chunk):
|
568 |
-
# try:
|
569 |
-
# sr, y = new_chunk[0], new_chunk[1]
|
570 |
-
# except TypeError:
|
571 |
-
# print(f"Error chunk structure: {type(new_chunk)}, content: {new_chunk}")
|
572 |
-
# return stream, "", None
|
573 |
-
|
574 |
-
# y = y.astype(np.float32) / np.max(np.abs(y))
|
575 |
-
|
576 |
-
# if stream is not None:
|
577 |
-
# stream = np.concatenate([stream, y])
|
578 |
-
# else:
|
579 |
-
# stream = y
|
580 |
-
|
581 |
-
# result = pipe_asr({"array": stream, "sampling_rate": sr}, return_timestamps=False)
|
582 |
-
|
583 |
-
# full_text = result.get("text","")
|
584 |
-
|
585 |
-
# return stream, full_text, result
|
586 |
-
|
587 |
def transcribe_function(stream, new_chunk):
|
588 |
try:
|
589 |
sr, y = new_chunk[0], new_chunk[1]
|
@@ -591,29 +534,16 @@ def transcribe_function(stream, new_chunk):
|
|
591 |
print(f"Error chunk structure: {type(new_chunk)}, content: {new_chunk}")
|
592 |
return stream, "", None
|
593 |
|
594 |
-
|
595 |
-
if y.size == 0:
|
596 |
-
print("Received an empty audio chunk, skipping processing.")
|
597 |
-
return stream, "", None
|
598 |
-
|
599 |
-
y = y.astype(np.float32)
|
600 |
-
|
601 |
-
# Check if y is all zeros or contains no valid data
|
602 |
-
if np.max(np.abs(y)) == 0:
|
603 |
-
print("Audio chunk contains only zeros, skipping normalization.")
|
604 |
-
normalized_y = y
|
605 |
-
else:
|
606 |
-
normalized_y = y / np.max(np.abs(y))
|
607 |
|
608 |
if stream is not None:
|
609 |
-
stream = np.concatenate([stream,
|
610 |
else:
|
611 |
-
stream =
|
612 |
|
613 |
-
# Placeholder for ASR pipeline; replace with actual pipeline
|
614 |
result = pipe_asr({"array": stream, "sampling_rate": sr}, return_timestamps=False)
|
615 |
|
616 |
-
full_text = result.get("text",
|
617 |
|
618 |
return stream, full_text, result
|
619 |
|
@@ -621,6 +551,8 @@ def transcribe_function(stream, new_chunk):
|
|
621 |
|
622 |
|
623 |
|
|
|
|
|
624 |
def update_map_with_response(history):
|
625 |
if not history:
|
626 |
return ""
|
|
|
346 |
|
347 |
|
348 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
349 |
def bot(history, choice, tts_choice, retrieval_mode):
|
350 |
if not history:
|
351 |
return history
|
|
|
379 |
|
380 |
history.append([response, None]) # Ensure the response is added in the correct format
|
381 |
|
382 |
+
|
383 |
+
|
|
|
|
|
384 |
|
385 |
|
386 |
|
|
|
527 |
|
528 |
#Normal Code with sample rate is 44100 Hz
|
529 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
530 |
def transcribe_function(stream, new_chunk):
|
531 |
try:
|
532 |
sr, y = new_chunk[0], new_chunk[1]
|
|
|
534 |
print(f"Error chunk structure: {type(new_chunk)}, content: {new_chunk}")
|
535 |
return stream, "", None
|
536 |
|
537 |
+
y = y.astype(np.float32) / np.max(np.abs(y))
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
538 |
|
539 |
if stream is not None:
|
540 |
+
stream = np.concatenate([stream, y])
|
541 |
else:
|
542 |
+
stream = y
|
543 |
|
|
|
544 |
result = pipe_asr({"array": stream, "sampling_rate": sr}, return_timestamps=False)
|
545 |
|
546 |
+
full_text = result.get("text","")
|
547 |
|
548 |
return stream, full_text, result
|
549 |
|
|
|
551 |
|
552 |
|
553 |
|
554 |
+
|
555 |
+
|
556 |
def update_map_with_response(history):
|
557 |
if not history:
|
558 |
return ""
|