Spaces:
Runtime error
Runtime error
File size: 2,648 Bytes
e698260 3273372 0d5492e e698260 f9f41df e698260 bef4887 e698260 8341dac bef4887 8341dac e698260 4198c04 e698260 1d55aac f9f41df e698260 3273372 e698260 4198c04 f9f41df e698260 1d55aac e698260 bdec318 532f375 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 |
from ui import *
from adjust import *
from set_up import prepare_input, prepare_video_subtitle
import gradio as gr
with gr.Blocks() as demo:
ui_lang_radio.render()
ui_lang_radio.change(change_lang, inputs=ui_lang_radio,
outputs=comp_to_update)
top_markdown.render()
with gr.Column():
with gr.Row():
with gr.Column():
input_url.render()
url_download_button.render()
url_download_button.click(
get_video_from_url, input_url, input_video)
input_video.render()
input_video.change(get_duration, input_video, [
start_time, end_time])
with gr.Column():
with gr.Row():
start_time.render()
end_time.render()
lang_radio.render()
model_dropdown.render()
# summary_checkbox.render()
start_button.render()
(
start_button
.click(prepare_input, [input_video, start_time, end_time, lang_radio, model_dropdown], [output_transcribe, output_file])
.then(prepare_output, inputs=input_video, outputs=[adjust_speaker, adjust_audio, prev_button, next_button, adjust_button])
.then(prepare_video_subtitle, inputs=[input_video, start_time, end_time], outputs=output_video)
)
bottom_markdown.render()
with gr.Row(equal_height=False):
with gr.Column():
output_video.render()
output_file.render()
with gr.Column():
output_transcribe.render()
# output_summary.render()
middle_markdown.render()
with gr.Row(equal_height=False):
adjust_audio.render()
adjust_speaker.render()
with gr.Row():
prev_button.render()
next_button.render()
prev_button.click(get_speakers_previous, inputs=[adjust_speaker], outputs=[
adjust_speaker, adjust_audio])
next_button.click(get_speakers_next, inputs=[adjust_speaker], outputs=[
adjust_speaker, adjust_audio])
adjust_button.render()
(
adjust_button
.click(start_adjust, inputs=[input_video, adjust_speaker], outputs=[output_transcribe, output_file])
.then(start_adjust_subtitle, inputs=[input_video, start_time, end_time], outputs=output_video)
)
with gr.Accordion("Copyright"):
gr.Markdown("Created with OpenAI Whisper and Huggingface")
if __name__ == "__main__":
demo.queue(concurrency_count=3).launch(share=False, show_error=True)
|