ysharma HF Staff commited on
Commit
7ed3c51
·
1 Parent(s): c43bf7c
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -116,7 +116,7 @@ def engine(text_input):
116
 
117
  app = gr.Interface(engine,
118
  gr.inputs.Textbox(lines=5, label="Input Text"),
119
- gr.outputs.Video(type=None, label='Final Merged video'), href='https://huggingface.co/flair/ner-english-ontonotes-large' target='_blank'
120
  description="Firstly, generates speech from input-text using facebook's fastspeech2-en-ljspeech from hub.<br>Then, takes the input-text and extracts the entities in it using Flair NER model from <a href='https://huggingface.co/flair/ner-english-ontonotes-large' target='_blank'>HF Hub</a>. <br>Then, generate images using <a href='https://huggingface.co/spaces/multimodalart/latentdiffusion' target='_blank'>Multimodalart Space</a> for every entity separately.<br>Creates a video by stringing all the entity-images together. <br>Lastly, Fuses the AI enerated audio and video together to create a coherent movie for you to watch. <br><br>A fun little app that lets you turn your text to video (well, in some ways atleast :) ). More the entities in your text, More time to build the output, More fun" ,
121
  examples=["On April 17th Sunday George celebrated Easter. He is staying at Empire State building with his parents. He is a citizen of Canada and speaks English and French fluently. His role model is former president Obama. He got 1000 dollar from his mother to visit Disney World and to buy new iPhone mobile. George likes watching Game of Thrones.", "April is the month of Easter weekend. Visit places like Statue of Liberty with friends. Take at least 200 dollars in cash with you. Use Android phone to find places in Newyork City."],
122
  title="Generate Video from Text",
 
116
 
117
  app = gr.Interface(engine,
118
  gr.inputs.Textbox(lines=5, label="Input Text"),
119
+ gr.outputs.Video(type=None, label='Final Merged video'),
120
  description="Firstly, generates speech from input-text using facebook's fastspeech2-en-ljspeech from hub.<br>Then, takes the input-text and extracts the entities in it using Flair NER model from <a href='https://huggingface.co/flair/ner-english-ontonotes-large' target='_blank'>HF Hub</a>. <br>Then, generate images using <a href='https://huggingface.co/spaces/multimodalart/latentdiffusion' target='_blank'>Multimodalart Space</a> for every entity separately.<br>Creates a video by stringing all the entity-images together. <br>Lastly, Fuses the AI enerated audio and video together to create a coherent movie for you to watch. <br><br>A fun little app that lets you turn your text to video (well, in some ways atleast :) ). More the entities in your text, More time to build the output, More fun" ,
121
  examples=["On April 17th Sunday George celebrated Easter. He is staying at Empire State building with his parents. He is a citizen of Canada and speaks English and French fluently. His role model is former president Obama. He got 1000 dollar from his mother to visit Disney World and to buy new iPhone mobile. George likes watching Game of Thrones.", "April is the month of Easter weekend. Visit places like Statue of Liberty with friends. Take at least 200 dollars in cash with you. Use Android phone to find places in Newyork City."],
122
  title="Generate Video from Text",