numz commited on
Commit
eea10fe
Β·
verified Β·
1 Parent(s): d014a86

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +77 -47
README.md CHANGED
@@ -35,12 +35,37 @@ It improves the quality of the lip-sync videos generated by the [Wav2Lip tool](h
35
  * [β˜• Support Wav2lip Studio](#-support-wav2lip-studio)
36
 
37
  ## πŸš€ Updates
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
38
  **2024.02.09 Spped Up Update (Standalone version only)**
39
  - πŸ‘¬ Clone voice: Add controls to manage the voice clone (See Usage section)
40
  - 🎏 translate video: Add features to translate panel to manage translation (See Usage section)
41
  - πŸ“Ί Add Trim feature: Add a feature to trim the video.
42
  - πŸ”‘ Automatic mask: Add a feature to automatically calculate the mask parameters (padding, dilate...). You can change parameters if needed.
43
  - πŸš€ Speed up processes : All processes are now faster, Analysis, Face Swap, Generation in High quality
 
44
 
45
  **2024.01.20 Major Update (Standalone version only)**
46
  - β™» Manage project: Add a feature to manage multiple project
@@ -86,11 +111,12 @@ It improves the quality of the lip-sync videos generated by the [Wav2Lip tool](h
86
  - πŸ“° Control debug
87
  - πŸ› Fix resize factor bug
88
 
89
- ## πŸ”— Requirements
90
 
91
- - FFmpeg : download it from the [official FFmpeg site](https://ffmpeg.org/download.html). Follow the instructions appropriate for your operating system, note ffmpeg have to be accessible from the command line.
92
- - Make sure ffmpeg is in your PATH environment variable. If not, add it to your PATH environment variable.
93
- 1. pyannote.audio:You need to agree to share your contact information to access pyannote models.
 
94
  To do so, go to both link:
95
  - [pyannote diarization-3.1 huggingface repository](https://huggingface.co/pyannote/speaker-diarization-3.1)
96
  - [pyannote segmentation-3.0 huggingface repository](https://huggingface.co/pyannote/segmentation-3.0)
@@ -110,10 +136,9 @@ set each field and click "Agree and access repository"
110
  }
111
  ```
112
 
113
- ## πŸ’» Installation
114
- 1. Install [python 3.10.11](https://www.python.org/downloads/release/python-31011/)
115
- 2. Install [git](https://git-scm.com/downloads)
116
- 3. Check ffmpeg, python, cuda and git installation
117
  ```bash
118
  python --version
119
  git --version
@@ -131,8 +156,14 @@ set each field and click "Agree and access repository"
131
  Cuda compilation tools, release 11.8, V11.8.89
132
  Build cuda_11.8.r11.8/compiler.31833905_0
133
  ```
 
 
 
 
 
 
134
 
135
- # Windows Users
136
  1. Install [Cuda 11.8](https://developer.nvidia.com/cuda-11-8-0-download-archive) if not ever done.
137
  ![Illustration](demo/cuda.png)
138
  2. Install [Visual Studio](https://visualstudio.microsoft.com/fr/downloads/). During the install, make sure to include the Python and C++ packages in visual studio installer.
@@ -147,9 +178,9 @@ set each field and click "Agree and access repository"
147
  ```
148
  4. double click on wav2lip-studio.bat, that will install the requirements and download the models
149
 
150
- # MACOS Users
151
 
152
- 1. Install python 3.9 and other requirements
153
  ```
154
  brew update
155
  brew install [email protected]
@@ -158,21 +189,15 @@ set each field and click "Agree and access repository"
158
  git-lfs install
159
  xcode-select --install
160
  ```
161
- 2. Unzip Wav2lipStudio zip achive in a folder
162
-
163
- 3. Install environnement and requirements
164
  ```
165
- cd /YourWav2lipStudioFolder
166
- /opt/homebrew/bin/python3.9 -m venv venv
167
- ./venv/bin/python3.9 -m pip install torch==2.1.2 torchvision==0.16.2 torchaudio==2.1.2
168
- ./venv/bin/python3.9 -m pip install -r requirements.txt
169
- ./venv/bin/python3.9 -m pip install transformers==4.33.2
170
- ./venv/bin/python3.9 -m pip install numpy==1.24.4
171
  ```
172
-
173
- 4. if It doesn't works or too long on pip install -r requirements.txt
174
 
175
  ```
 
 
176
  ./venv/bin/python3.9 -m pip install inaSpeechSegmenter
177
  ./venv/bin/python3.9 -m pip install tyro==0.8.5 pykalman==0.9.7
178
  ./venv/bin/python3.9 -m pip install TTS==0.21.2
@@ -181,28 +206,29 @@ set each field and click "Agree and access repository"
181
  ./venv/bin/python3.9 -m pip install transformers==4.33.2
182
  ./venv/bin/python3.9 -m pip install numpy==1.24.4
183
  ```
184
- 4.1. for silicon
185
- ```
186
- ./venv/bin/python3.9 -m pip uninstall torch torchvision torchaudio
187
- ./venv/bin/python3.9 -m pip install --pre torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/nightly/cpu
188
- sed -i '' 's/from torchvision.transforms.functional_tensor import rgb_to_grayscale/from torchvision.transforms.functional import rgb_to_grayscale/' venv/lib/python3.9/site-packages/basicsr/data/degradations.py
189
- ```
190
- 5. Install models
 
191
  ```
192
  git clone https://huggingface.co/numz/wav2lip_studio-0.2 models
193
  git clone https://huggingface.co/KwaiVGI/LivePortrait models/pretrained_weights
194
  ```
195
- 6. Launch UI
196
  ```
197
  mkdir projects
198
  ./venv/bin/python3.9 wav2lip_studio.py
199
  ```
200
 
201
- ## Tutorial
202
  - [FR version](https://youtu.be/43Q8YASkcUA)
203
  - [EN Version](https://youtu.be/B84A5alpPDc)
204
 
205
- ## 🐍 Usage
206
  ##PARAMETERS
207
  1. Enter project name and click enter.
208
  2. Choose a video (avi or mp4 format). Note avi file will not appear in Video input but process will works.
@@ -243,20 +269,24 @@ set each field and click "Agree and access repository"
243
  ]
244
  ```
245
  3. Input Video: Allow to use audio from the input video, voices cloning and translation. see [Input Video](#input-video) section for more details.
246
- 11. **Video Quality**:
 
 
 
 
247
  - **Low**: Original Wav2Lip quality, fast but not very good.
248
  - **Medium**: Better quality by apply post processing on the mouth, slower.
249
  - **High**: Better quality by apply post processing and upscale the mouth quality, slower.
250
- 12. **Wav2lip Checkpoint**: Choose beetwen 2 wav2lip model:
251
  - **Wav2lip**: Original Wav2Lip model, fast but not very good.
252
  - **Wav2lip GAN**: Better quality by apply post processing on the mouth, slower.
253
- 13. **Face Restoration Model**: Choose beetwen 2 face restoration model:
254
  - **Code Former**:
255
  - A value of 0 offers higher quality but may significantly alter the person's facial appearance and cause noticeable flickering between frames.
256
  - A value of 1 provides lower quality but maintains the person's face more consistently and reduces frame flickering.
257
  - Using a value below 0.5 is not advised. Adjust this setting to achieve optimal results. Starting with a value of 0.75 is recommended.
258
  - **GFPGAN**: Usually better quality.
259
- 14. **Volume Amplifier**: Not amplify the volume of the output audio but allows you to amplify the volume of the audio when sending it to Wav2Lip. This allows you to better control on lips movement.
260
 
261
  ## KEYFRAMES MANAGER
262
  ![Illustration](demo/keyframes-manager.png)
@@ -319,7 +349,7 @@ For each segment of the translated text, you can :
319
  - Delete the segment by click on the trash button.
320
  - Add a new segment under this one by click on the arrow down button.
321
 
322
- ## πŸ“Ί Examples
323
 
324
  https://user-images.githubusercontent.com/800903/262439441-bb9d888a-d33e-4246-9f0a-1ddeac062d35.mp4
325
 
@@ -329,7 +359,7 @@ https://user-images.githubusercontent.com/800903/262449305-901086a3-22cb-42d2-b5
329
 
330
  https://user-images.githubusercontent.com/800903/267808494-300f8cc3-9136-4810-86e2-92f2114a5f9a.mp4
331
 
332
- ## πŸ“– Behind the scenes
333
 
334
  This extension operates in several stages to improve the quality of Wav2Lip-generated videos:
335
 
@@ -339,7 +369,7 @@ This extension operates in several stages to improve the quality of Wav2Lip-gene
339
  4. **Mask Creation**: The script creates a mask around the mouth and tries to keep other facial motions like those of the cheeks and chin.
340
  5. **Video Generation**: The script then takes the high-quality mouth image and overlays it onto the original image guided by the mouth mask.
341
 
342
- ## πŸ’ͺ Quality tips
343
  - Use a high quality video as input
344
  - Use a video with a consistent frame rate. Occasionally, videos may exhibit unusual playback frame rates (not the standard 24, 25, 30, 60), which can lead to issues with the face mask.
345
  - Use a high quality audio file as input, without background noise or music. Clean audio with a tool like [https://podcast.adobe.com/enhance](https://podcast.adobe.com/enhance).
@@ -347,12 +377,12 @@ This extension operates in several stages to improve the quality of Wav2Lip-gene
347
  - Mask Blur maximum twice the value of Mouth Mask Dilate. If you want to increase the blur, increase the value of Mouth Mask Dilate otherwise the mouth will be blurred and the underlying mouth could be visible.
348
  - Upscaling can be good for improving result, particularly around the mouth area. However, it will extend the processing duration. Use this tutorial from Olivio Sarikas to upscale your video: [https://www.youtube.com/watch?v=3z4MKUqFEUk](https://www.youtube.com/watch?v=3z4MKUqFEUk). Ensure the denoising strength is set between 0.0 and 0.05, select the 'revAnimated' model, and use the batch mode. i'll create a tutorial for this soon.
349
 
350
- ## ⚠ Noted Constraints
351
  - for speed up process try to keep resolution under 1000x1000px and upscaling after process.
352
  - If the initial phase is excessively lengthy, consider using the "resize factor" to decrease the video's dimensions.
353
  - While there's no strict size limit for videos, larger videos will require more processing time. It's advisable to employ the "resize factor" to minimize the video size and then upscale the video once processing is complete.
354
 
355
- ## know issues:
356
  If you have issues to install insightface, follow this step:
357
  - Download [insightface precompiled](https://github.com/Gourieff/Assets/raw/main/Insightface/insightface-0.7.3-cp310-cp310-win_amd64.whl) and paste it in the root folder of Wav2lip-studio
358
  - in terminal go to wav2lip-studio folder and type the following commands:
@@ -363,25 +393,25 @@ python -m pip install insightface-0.7.3-cp310-cp310-win_amd64.whl
363
  ```
364
  Enjoy
365
 
366
- ## πŸ“ To do
367
  - βœ”οΈ Standalone version
368
  - βœ”οΈ Add a way to use a face swap image
369
  - βœ”οΈ Add Possibility to use a video for audio input
370
  - βœ”οΈ Convert avi to mp4. Avi is not show in video input but process work fine
371
  - [ ] ComfyUI intergration
372
 
373
- ## 😎 Contributing
374
 
375
  We welcome contributions to this project. When submitting pull requests, please provide a detailed description of the changes. see [CONTRIBUTING](CONTRIBUTING.md) for more information.
376
 
377
- ## πŸ™ Appreciation
378
  - [Wav2Lip](https://github.com/Rudrabha/Wav2Lip)
379
  - [CodeFormer](https://github.com/sczhou/CodeFormer)
380
  - [Coqui TTS](https://github.com/coqui-ai/TTS)
381
  - [facefusion](https://github.com/facefusion/facefusion)
382
  - [Vocal Remover](https://github.com/tsurumeso/vocal-remover)
383
 
384
- ## β˜• Support Wav2lip Studio
385
 
386
  this project is open-source effort that is free to use and modify. I rely on the support of users to keep this project going and help improve it. If you'd like to support me, you can make a donation on my Patreon page. Any contribution, large or small, is greatly appreciated!
387
 
@@ -389,7 +419,7 @@ Your support helps me cover the costs of development and maintenance, and allows
389
 
390
  [patreon page](https://www.patreon.com/Wav2LipStudio)
391
 
392
- ## πŸ“ Citation
393
  If you use this project in your own work, in articles, tutorials, or presentations, we encourage you to cite this project to acknowledge the efforts put into it.
394
 
395
  To cite this project, please use the following BibTeX format:
@@ -405,5 +435,5 @@ To cite this project, please use the following BibTeX format:
405
  }
406
  ```
407
 
408
- ## πŸ“œ License
409
  * The code in this repository is released under the MIT license as found in the [LICENSE file](LICENSE).
 
35
  * [β˜• Support Wav2lip Studio](#-support-wav2lip-studio)
36
 
37
  ## πŸš€ Updates
38
+ **2024.10.13 Add avatar for driving video**
39
+ - πŸ’ͺ Add 10 new avatars for driving video, you can now choose an avatar before generate the driving video.
40
+ - πŸ“Ί Add a feature to close or not the mouth before generating lip sync video.
41
+ - πŸ› Easy docker installation, follow instructions bellow.
42
+ - β™» Better macos integration, follow instructions bellow.
43
+ - πŸš€ In Comfyui panel, you can now regenerate mask and keyframe after modification of your video, allow better mouth mask.
44
+
45
+ **2024.09.03 ComfyUI Integration in Lip Sync Studio**
46
+ - πŸ’ͺManage and chain your comfyui worklows from end to end.
47
+
48
+ **2024.08.07 Major Update (Standalone version only)**
49
+ - πŸ“Ί"Add Driving video feature": this feature allows you to generate a driving video to generate better lip sync.
50
+
51
+ **2024.05.06 Major Update (Standalone version only)**
52
+ - πŸ›"Data Structure": I had to restructure the files to allow for better quality in the video output. The previous version did everything in RAM at the expense of video quality; each pass degraded the videos, for example, if you did a face swap + Wav2Lip, there was a degradation of quality because of creating a first pass for Wav2Lip and a second for face swap. You will now find a "data" directory in each project containing all the files necessary for the tool's work and maintaining quality through different passes (quality above all).
53
+ - β™»"Wav2Lip Video Outputs": After generating Wav2Lip videos, the videos are numbered in the output directory. Clicking on "video quality" loads the last video of the specified quality.
54
+ - πŸ‘„"Zero Mouth": this feature should allow closing a person's mouth before proceeding with lip-syncing, sometimes it doesn't have much effect or can add some flickering to the image, but I have had good results in some cases. Technically, this will take two passes to close the mouth, you will find the frames used by the tool in "data\zero."
55
+ - πŸ‘¬"Clone Voice": the interface has been revised.
56
+ - πŸ’ͺ"High Quality Vs Best Quality": In this version, there is not much difference between High and Best. Best is to be used with videos where faces are large on the screen like on a 4K video for example. The process behind just uses different GFPGAN models and a different face alignment.
57
+ - β–Ά "Show Frame Number": In Low Quality only, the frame number appears in the top left corner. This helps to identify the frame where you want to make modifications.
58
+ - πŸ“Ί"Show Wav2Lip Output": this feature allows you to see the Wav2Lip output taking into account the input audio.
59
+ - "New Face Alignment": The face alignment has been reviewed.
60
+ - πŸ”‘"Zoom In, Zoom Out, Move Right,...": Now you will understand why sometimes the results are strange and generate deformed lips, broken teeth, or other very strange things.I recommend the video tutorial here: https://www.patreon.com/posts/key-feature-103716855
61
+
62
  **2024.02.09 Spped Up Update (Standalone version only)**
63
  - πŸ‘¬ Clone voice: Add controls to manage the voice clone (See Usage section)
64
  - 🎏 translate video: Add features to translate panel to manage translation (See Usage section)
65
  - πŸ“Ί Add Trim feature: Add a feature to trim the video.
66
  - πŸ”‘ Automatic mask: Add a feature to automatically calculate the mask parameters (padding, dilate...). You can change parameters if needed.
67
  - πŸš€ Speed up processes : All processes are now faster, Analysis, Face Swap, Generation in High quality
68
+ - πŸ’ͺ Less disk space used : Remove temporary files after generation and keep only necessary data, will greatly reduce disk space used.
69
 
70
  **2024.01.20 Major Update (Standalone version only)**
71
  - β™» Manage project: Add a feature to manage multiple project
 
111
  - πŸ“° Control debug
112
  - πŸ› Fix resize factor bug
113
 
114
+ # πŸ’» Installation
115
 
116
+ ## πŸ”— Requirements (windows, linux, macos)
117
+ 1. FFmpeg : download it from the [official FFmpeg site](https://ffmpeg.org/download.html). Follow the instructions appropriate for your operating system, note ffmpeg have to be accessible from the command line.
118
+ - Make sure ffmpeg is in your PATH environment variable. If not, add it to your PATH environment variable.
119
+ 2. pyannote.audio:You need to agree to share your contact information to access pyannote models.
120
  To do so, go to both link:
121
  - [pyannote diarization-3.1 huggingface repository](https://huggingface.co/pyannote/speaker-diarization-3.1)
122
  - [pyannote segmentation-3.0 huggingface repository](https://huggingface.co/pyannote/segmentation-3.0)
 
136
  }
137
  ```
138
 
139
+ 3. Install [python 3.10.11](https://www.python.org/downloads/release/python-31011/) (for mac users follow instructions bellow)
140
+ 4. Install [git](https://git-scm.com/downloads)
141
+ 5. Check ffmpeg, python, cuda and git installation
 
142
  ```bash
143
  python --version
144
  git --version
 
156
  Cuda compilation tools, release 11.8, V11.8.89
157
  Build cuda_11.8.r11.8/compiler.31833905_0
158
  ```
159
+
160
+ ## Linux Users
161
+ 1. make sure to have git-lfs installed
162
+ ```bash
163
+ sudo apt-get install git-lfs
164
+ ```
165
 
166
+ ## Windows Users
167
  1. Install [Cuda 11.8](https://developer.nvidia.com/cuda-11-8-0-download-archive) if not ever done.
168
  ![Illustration](demo/cuda.png)
169
  2. Install [Visual Studio](https://visualstudio.microsoft.com/fr/downloads/). During the install, make sure to include the Python and C++ packages in visual studio installer.
 
178
  ```
179
  4. double click on wav2lip-studio.bat, that will install the requirements and download the models
180
 
181
+ ## MACOS Users
182
 
183
+ 1. Install python 3.9
184
  ```
185
  brew update
186
  brew install [email protected]
 
189
  git-lfs install
190
  xcode-select --install
191
  ```
192
+ 2. unzip the wav2lip-studio.zip in a folder
 
 
193
  ```
194
+ unzip wav2lip-studio.zip
 
 
 
 
 
195
  ```
196
+ 3. Install environnement and requirements
 
197
 
198
  ```
199
+ cd /YourWav2lipStudioFolder
200
+ /opt/homebrew/bin/python3.9 -m venv venv
201
  ./venv/bin/python3.9 -m pip install inaSpeechSegmenter
202
  ./venv/bin/python3.9 -m pip install tyro==0.8.5 pykalman==0.9.7
203
  ./venv/bin/python3.9 -m pip install TTS==0.21.2
 
206
  ./venv/bin/python3.9 -m pip install transformers==4.33.2
207
  ./venv/bin/python3.9 -m pip install numpy==1.24.4
208
  ```
209
+ 3.1. for silicon, one more step is needed
210
+ ```
211
+ ./venv/bin/python3.9 -m pip uninstall torch torchvision torchaudio
212
+ ./venv/bin/python3.9 -m pip install --pre torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/nightly/cpu
213
+ sed -i '' 's/from torchvision.transforms.functional_tensor import rgb_to_grayscale/from torchvision.transforms.functional import rgb_to_grayscale/' venv/lib/python3.9/site-packages/basicsr/data/degradations.py
214
+ ```
215
+
216
+ 4. Install models
217
  ```
218
  git clone https://huggingface.co/numz/wav2lip_studio-0.2 models
219
  git clone https://huggingface.co/KwaiVGI/LivePortrait models/pretrained_weights
220
  ```
221
+ 5. Launch UI
222
  ```
223
  mkdir projects
224
  ./venv/bin/python3.9 wav2lip_studio.py
225
  ```
226
 
227
+ # Tutorial
228
  - [FR version](https://youtu.be/43Q8YASkcUA)
229
  - [EN Version](https://youtu.be/B84A5alpPDc)
230
 
231
+ # 🐍 Usage
232
  ##PARAMETERS
233
  1. Enter project name and click enter.
234
  2. Choose a video (avi or mp4 format). Note avi file will not appear in Video input but process will works.
 
269
  ]
270
  ```
271
  3. Input Video: Allow to use audio from the input video, voices cloning and translation. see [Input Video](#input-video) section for more details.
272
+ 11. **Driving Video**: Choose an avatar to generate a driving video.
273
+ - **Avatars**: Choose between 10 avatars to use for the driving video, each one will give a different driving result on lipsync output video.
274
+ - **Close Mouth**: Close the mouth of the avatar before generating the driving video.
275
+ - **Generate Driving Video**: Generate the driving video.
276
+ 12. **Video Quality**:
277
  - **Low**: Original Wav2Lip quality, fast but not very good.
278
  - **Medium**: Better quality by apply post processing on the mouth, slower.
279
  - **High**: Better quality by apply post processing and upscale the mouth quality, slower.
280
+ 13. **Wav2lip Checkpoint**: Choose beetwen 2 wav2lip model:
281
  - **Wav2lip**: Original Wav2Lip model, fast but not very good.
282
  - **Wav2lip GAN**: Better quality by apply post processing on the mouth, slower.
283
+ 14. **Face Restoration Model**: Choose beetwen 2 face restoration model:
284
  - **Code Former**:
285
  - A value of 0 offers higher quality but may significantly alter the person's facial appearance and cause noticeable flickering between frames.
286
  - A value of 1 provides lower quality but maintains the person's face more consistently and reduces frame flickering.
287
  - Using a value below 0.5 is not advised. Adjust this setting to achieve optimal results. Starting with a value of 0.75 is recommended.
288
  - **GFPGAN**: Usually better quality.
289
+ 15. **Volume Amplifier**: Not amplify the volume of the output audio but allows you to amplify the volume of the audio when sending it to Wav2Lip. This allows you to better control on lips movement.
290
 
291
  ## KEYFRAMES MANAGER
292
  ![Illustration](demo/keyframes-manager.png)
 
349
  - Delete the segment by click on the trash button.
350
  - Add a new segment under this one by click on the arrow down button.
351
 
352
+ # πŸ“Ί Examples
353
 
354
  https://user-images.githubusercontent.com/800903/262439441-bb9d888a-d33e-4246-9f0a-1ddeac062d35.mp4
355
 
 
359
 
360
  https://user-images.githubusercontent.com/800903/267808494-300f8cc3-9136-4810-86e2-92f2114a5f9a.mp4
361
 
362
+ # πŸ“– Behind the scenes
363
 
364
  This extension operates in several stages to improve the quality of Wav2Lip-generated videos:
365
 
 
369
  4. **Mask Creation**: The script creates a mask around the mouth and tries to keep other facial motions like those of the cheeks and chin.
370
  5. **Video Generation**: The script then takes the high-quality mouth image and overlays it onto the original image guided by the mouth mask.
371
 
372
+ # πŸ’ͺ Quality tips
373
  - Use a high quality video as input
374
  - Use a video with a consistent frame rate. Occasionally, videos may exhibit unusual playback frame rates (not the standard 24, 25, 30, 60), which can lead to issues with the face mask.
375
  - Use a high quality audio file as input, without background noise or music. Clean audio with a tool like [https://podcast.adobe.com/enhance](https://podcast.adobe.com/enhance).
 
377
  - Mask Blur maximum twice the value of Mouth Mask Dilate. If you want to increase the blur, increase the value of Mouth Mask Dilate otherwise the mouth will be blurred and the underlying mouth could be visible.
378
  - Upscaling can be good for improving result, particularly around the mouth area. However, it will extend the processing duration. Use this tutorial from Olivio Sarikas to upscale your video: [https://www.youtube.com/watch?v=3z4MKUqFEUk](https://www.youtube.com/watch?v=3z4MKUqFEUk). Ensure the denoising strength is set between 0.0 and 0.05, select the 'revAnimated' model, and use the batch mode. i'll create a tutorial for this soon.
379
 
380
+ # ⚠ Noted Constraints
381
  - for speed up process try to keep resolution under 1000x1000px and upscaling after process.
382
  - If the initial phase is excessively lengthy, consider using the "resize factor" to decrease the video's dimensions.
383
  - While there's no strict size limit for videos, larger videos will require more processing time. It's advisable to employ the "resize factor" to minimize the video size and then upscale the video once processing is complete.
384
 
385
+ # know issues:
386
  If you have issues to install insightface, follow this step:
387
  - Download [insightface precompiled](https://github.com/Gourieff/Assets/raw/main/Insightface/insightface-0.7.3-cp310-cp310-win_amd64.whl) and paste it in the root folder of Wav2lip-studio
388
  - in terminal go to wav2lip-studio folder and type the following commands:
 
393
  ```
394
  Enjoy
395
 
396
+ # πŸ“ To do
397
  - βœ”οΈ Standalone version
398
  - βœ”οΈ Add a way to use a face swap image
399
  - βœ”οΈ Add Possibility to use a video for audio input
400
  - βœ”οΈ Convert avi to mp4. Avi is not show in video input but process work fine
401
  - [ ] ComfyUI intergration
402
 
403
+ # 😎 Contributing
404
 
405
  We welcome contributions to this project. When submitting pull requests, please provide a detailed description of the changes. see [CONTRIBUTING](CONTRIBUTING.md) for more information.
406
 
407
+ # πŸ™ Appreciation
408
  - [Wav2Lip](https://github.com/Rudrabha/Wav2Lip)
409
  - [CodeFormer](https://github.com/sczhou/CodeFormer)
410
  - [Coqui TTS](https://github.com/coqui-ai/TTS)
411
  - [facefusion](https://github.com/facefusion/facefusion)
412
  - [Vocal Remover](https://github.com/tsurumeso/vocal-remover)
413
 
414
+ # β˜• Support Wav2lip Studio
415
 
416
  this project is open-source effort that is free to use and modify. I rely on the support of users to keep this project going and help improve it. If you'd like to support me, you can make a donation on my Patreon page. Any contribution, large or small, is greatly appreciated!
417
 
 
419
 
420
  [patreon page](https://www.patreon.com/Wav2LipStudio)
421
 
422
+ # πŸ“ Citation
423
  If you use this project in your own work, in articles, tutorials, or presentations, we encourage you to cite this project to acknowledge the efforts put into it.
424
 
425
  To cite this project, please use the following BibTeX format:
 
435
  }
436
  ```
437
 
438
+ # πŸ“œ License
439
  * The code in this repository is released under the MIT license as found in the [LICENSE file](LICENSE).