Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
import gradio as gr
|
2 |
import torch
|
3 |
-
|
4 |
from PIL import Image, ImageDraw, ImageFont
|
5 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
6 |
from min_dalle import MinDalle
|
@@ -214,9 +214,9 @@ def get_output_video(text):
|
|
214 |
for m in file_names:
|
215 |
duration = mp3_lengths[d]
|
216 |
print(d, duration)
|
217 |
-
clips.append(ImageClip(m).set_duration(duration + 0.5))
|
218 |
d += 1
|
219 |
-
concat_clip = concatenate_videoclips(clips, method="compose")
|
220 |
concat_clip.write_videofile("result_new.mp4", fps=24)
|
221 |
|
222 |
# Step 10 - Merge Video + Audio
|
@@ -225,7 +225,6 @@ def get_output_video(text):
|
|
225 |
movie_final = 'result_final.mp4'
|
226 |
|
227 |
def combine_audio(vidname, audname, outname, fps=24):
|
228 |
-
import moviepy.editor as mpe
|
229 |
my_clip = mpe.VideoFileClip(vidname)
|
230 |
audio_background = mpe.AudioFileClip(audname)
|
231 |
final_clip = my_clip.set_audio(audio_background)
|
|
|
1 |
import gradio as gr
|
2 |
import torch
|
3 |
+
import moviepy.editor as mpe
|
4 |
from PIL import Image, ImageDraw, ImageFont
|
5 |
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM
|
6 |
from min_dalle import MinDalle
|
|
|
214 |
for m in file_names:
|
215 |
duration = mp3_lengths[d]
|
216 |
print(d, duration)
|
217 |
+
clips.append(mpe.ImageClip(m).set_duration(duration + 0.5))
|
218 |
d += 1
|
219 |
+
concat_clip = mpe.concatenate_videoclips(clips, method="compose")
|
220 |
concat_clip.write_videofile("result_new.mp4", fps=24)
|
221 |
|
222 |
# Step 10 - Merge Video + Audio
|
|
|
225 |
movie_final = 'result_final.mp4'
|
226 |
|
227 |
def combine_audio(vidname, audname, outname, fps=24):
|
|
|
228 |
my_clip = mpe.VideoFileClip(vidname)
|
229 |
audio_background = mpe.AudioFileClip(audname)
|
230 |
final_clip = my_clip.set_audio(audio_background)
|