File size: 9,332 Bytes
11cf479
486b9c4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
d627392
486b9c4
 
 
 
 
 
 
d627392
486b9c4
 
8aca7f6
486b9c4
 
d627392
486b9c4
 
 
 
 
49487f1
486b9c4
 
 
 
 
 
 
 
 
 
c3dd560
 
 
 
 
486b9c4
 
 
 
 
 
 
 
 
 
 
 
c3dd560
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6695749
bd2f353
486b9c4
 
 
 
 
 
 
 
 
 
 
 
 
bd2f353
 
486b9c4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
49487f1
486b9c4
 
49487f1
b951d0b
49487f1
 
486b9c4
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
bd2f353
486b9c4
 
 
 
 
8a67596
d627392
486b9c4
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
import os
import subprocess
import sys
import uuid
import gradio as gr
from pydub import AudioSegment
from TTS.api import TTS

# Инициализация моделей TTS
tts = TTS("tts_models/multilingual/multi-dataset/xtts_v2", )
# tts.to("cuda")

# Опции языков
language_options = {
    "English (en)": "en",
    "Spanish (es)": "es",
    "French (fr)": "fr",
    "German (de)": "de",
    "Italian (it)": "it",
    "Portuguese (pt)": "pt",
    "Polish (pl)": "pl",
    "Turkish (tr)": "tr",
    "Russian (ru)": "ru",
    "Dutch (nl)": "nl",
    "Czech (cs)": "cs",
    "Arabic (ar)": "ar",
    "Chinese (zh-cn)": "zh-cn",
    "Japanese (ja)": "ja",
    "Hungarian (hu)": "hu",
    "Korean (ko)": "ko",
    "Hindi (hi)": "hi"
}

other_language = {
    "Vietnamese": "vie",
    "Serbian": "srp",
    "Romanian": "ron",
    "Indonesian": "ind",
    "Philippine": "tgl"
}

def clean_audio(audio_path):
    out_filename = f"output/cleaned_{uuid.uuid4()}.wav"
    lowpass_highpass = "lowpass=8000,highpass=75,"
    trim_silence = "areverse,silenceremove=start_periods=1:start_silence=0:start_threshold=0.02,areverse,silenceremove=start_periods=1:start_silence=0:start_threshold=0.02,"
    try:
        shell_command = f"ffmpeg -y -i {audio_path} -af {lowpass_highpass}{trim_silence} {out_filename}".split()
        subprocess.run(shell_command, capture_output=True, check=True)
        print(f"Audio cleaned and saved to {out_filename}")
        return out_filename
    except subprocess.CalledProcessError as e:
        print(f"Error during audio cleaning: {e}")
        return audio_path

def check_audio_length(audio_path, max_duration=120):
    try:
        audio = AudioSegment.from_file(audio_path)
        duration = audio.duration_seconds
        if duration > max_duration:
            print(f"Audio is too long: {duration} seconds. Max allowed is {max_duration} seconds.")
            return False
        return True
    except Exception as e:
        print(f"Error while checking audio length: {e}")
        return False

def synthesize_and_convert_voice(text, language_iso, voice_audio_path, speed):
    tts_synthesis = TTS(model_name=f"tts_models/{language_iso}/fairseq/vits", )
    wav_data = tts_synthesis.tts(text, speed=speed)
    tts_conversion = TTS(model_name="voice_conversion_models/multilingual/vctk/freevc24", progress_bar=False)
    output_file = "output/docout.wav"
    os.makedirs("output", exist_ok=True)
    tts_conversion.voice_conversion_to_file(wav_data, target_wav=voice_audio_path,
                                            file_path=output_file)
    return output_file  # Возвращаем путь к сгенерированному аудио

def synthesize_speech(text, speaker_wav_path, language_iso, speed):
    output_file_xtts = "output/undocout.wav"
    tts.tts_to_file(text=text, file_path=output_file_xtts, speed=speed, speaker_wav=speaker_wav_path,
                    language=language_iso)
    tts_conversion = TTS(model_name="voice_conversion_models/multilingual/vctk/freevc24", progress_bar=False)
    output_file = "output/docout.wav"
    os.makedirs("output", exist_ok=True)
    tts_conversion.voice_conversion_to_file(output_file_xtts, target_wav=speaker_wav_path,
                                            file_path=output_file)
    return output_file  # Возвращаем путь к сгенерированному аудио

def get_language_code(selected_language):
    if selected_language in language_options:
        return language_options[selected_language]
    elif selected_language in other_language:
        return other_language[selected_language]
    else:
        return None

def process_speech(text, speaker_wav, selected_language, speed):
    language_code = get_language_code(selected_language)

    if language_code is None:
        raise ValueError("Выбранный язык не поддерживается.")

    # Проверка длины аудио
    if not check_audio_length(speaker_wav):
        error_message = "Длина аудио превышает допустимый лимит в 2 минуты."
        error = gr.Error(error_message, duration=5)
        raise error

    cleaned_wav_path = clean_audio(speaker_wav)

    if selected_language in other_language:
        audio_path = synthesize_and_convert_voice(text, language_code, cleaned_wav_path, speed)
    else:
        audio_path = synthesize_speech(text, cleaned_wav_path, language_code, speed)

    return audio_path

def generate_lipsync(video_path, audio_path, pad_top, pad_bottom, pad_left, pad_right, no_smooth, save_as_video):
    output_dir = "outputs"
    os.makedirs(output_dir, exist_ok=True)
    output_path = os.path.join(output_dir, "output.mp4")

    args = [
        "--checkpoint_path", "checkpoints/wav2lip_gan.pth",
        "--segmentation_path", "checkpoints/face_segmentation.pth",    
        "--no_seg",  
        "--no_sr",      
        "--face", video_path,
        "--audio", audio_path,
        "--outfile", output_path,
        "--resize_factor", "2", 
        "--face_det_batch_size", "4",    
        "--wav2lip_batch_size", "64",    
        "--fps", "30",
        "--pads", str(pad_top), str(pad_bottom), str(pad_left), str(pad_right)
    ]

    if no_smooth:
        args.append("--nosmooth")
    if save_as_video:
        args.append("--save_as_video")

    try:
        cmd = ["python", "inference.py"] + args
        print(f"Запуск инференса с командой: {' '.join(cmd)}")
        subprocess.run(cmd, check=True)
    except subprocess.CalledProcessError as e:
        print(f"Ошибка при выполнении команды: {e}")
        return f"Произошла ошибка при обработке: {e}"

    if not os.path.exists(output_path):
        print("Выходной файл не существует.")
        return "Не удалось создать выходное видео."

    print(f"Выходной файл создан по пути: {output_path}")
    return output_path  

def process_all(text, speaker_wav, selected_language, speed, video, pad_top, pad_bottom, pad_left, pad_right, no_smooth, save_as_video):
    # Шаг 1: Генерация аудио с клонированным голосом
    audio_path = process_speech(text, speaker_wav, selected_language, speed)

    # Шаг 2: Генерация видео с липсинком
    video_path = video  # Предполагается, что video — это путь к файлу

    result = generate_lipsync(video_path, audio_path, pad_top, pad_bottom, pad_left, pad_right, no_smooth, save_as_video)
    return result

with gr.Blocks() as demo:
    gr.Markdown("# Объединение Voice Clone и Lipsync")

    with gr.Row():
        with gr.Column():
            gr.Markdown("### Шаг 1: Настройки синтеза речи")
            text_input = gr.Textbox(label="Введите текст для генерации", placeholder="Введите ваш текст здесь...")
            speaker_wav_input = gr.Audio(label="Загрузите аудио говорящего (WAV формат)", type="filepath")

            all_languages = list(language_options.keys()) + list(other_language.keys())
            language_input = gr.Dropdown(
                label="Язык",
                choices=all_languages,
                value="English (en)"
            )

            speed_input = gr.Slider(
                label="Скорость синтеза",
                minimum=0.1,
                maximum=10,
                step=0.1,
                value=1.0,
                info="Выберите скорость"
            )

        with gr.Column():
            gr.Markdown("### Шаг 2: Настройки липсинка")
            video_input = gr.File(label="Видео или Изображение", type="filepath")
            pad_top = gr.Slider(minimum=0, maximum=50, step=1, value=0, label="Отступ сверху")
            pad_bottom = gr.Slider(minimum=0, maximum=50, step=1, value=10, label="Отступ снизу")
            pad_left = gr.Slider(minimum=0, maximum=50, step=1, value=0, label="Отступ слева")
            pad_right = gr.Slider(minimum=0, maximum=50, step=1, value=0, label="Отступ справа")
            no_smooth = gr.Checkbox(label="Без сглаживания", value=False)
            save_as_video = gr.Checkbox(label="Сохранять как видео", value=True)

    output_video = gr.Video(label="Сгенерированное видео")

    with gr.Row():
        generate_button = gr.Button("Сгенерировать")
        gr.HTML("<div style='width:300px;'></div>")
        reload_button = gr.Button("Перезапустить")

    generate_button.click(
        fn=process_all,
        inputs=[text_input, speaker_wav_input, language_input, speed_input, video_input, pad_top, pad_bottom, pad_left, pad_right, no_smooth, save_as_video],
        outputs=output_video
    )

    reload_button.click(fn=lambda: os._exit(0), inputs=None, outputs=None)

def launch_gradio():
    demo.launch(
        debug=True,
        server_port=8600,
        server_name="0.0.0.0",
    )

if __name__ == "__main__":
    launch_gradio()