File size: 4,850 Bytes
4068b97 027b663 4068b97 a4b9b41 4068b97 1a492ad 4068b97 027b663 4068b97 1a492ad 0af3599 4068b97 1a492ad 1e1edb9 1a492ad 1e1edb9 1a492ad 1e1edb9 1a492ad 1e1edb9 4068b97 1a492ad 4068b97 1a492ad 4068b97 1a492ad 4068b97 1a492ad 4068b97 1a492ad 4068b97 1a492ad 027b663 1a492ad 148f5fc 1a492ad 1aab015 4068b97 1a492ad 4068b97 027b663 2a3fdcb 4068b97 027b663 0af3599 1a492ad 027b663 0af3599 148f5fc 027b663 81dc6a9 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 |
import os
import numpy as np
import tempfile
from scipy.io import wavfile
import gradio as gr
from inference import EnsembleDemucsMDXMusicSeparationModel, predict_with_model
import torch
import time
# 检查文件是否准备好
def check_file_readiness(filepath):
num_same_size_checks = 0
last_size = -1
while num_same_size_checks < 5:
current_size = os.path.getsize(filepath)
if current_size == last_size:
num_same_size_checks += 1
else:
num_same_size_checks = 0
last_size = current_size
time.sleep(0.5)
return True
# 音乐分离的包装函数
def separate_music_file_wrapper(uploaded_files, use_cpu, use_single_onnx, large_overlap, small_overlap, chunk_size, use_large_gpu):
input_files = [uploaded_files.name] # 获取上传文件的路径
# 验证重叠值
if not (0 <= large_overlap <= 1) or not (0 <= small_overlap <= 1):
raise ValueError("重叠值必须在0到1之间。")
# 验证块大小
if chunk_size <= 0:
raise ValueError("块大小必须大于0。") # 太小了
options = {
'input_audio': input_files,
'output_folder': 'results',
'cpu': use_cpu,
'single_onnx': use_single_onnx,
'overlap_large': large_overlap,
'overlap_small': small_overlap,
'chunk_size': chunk_size,
'large_gpu': use_large_gpu,
}
predict_with_model(options)
# 清除GPU缓存
if torch.cuda.is_available():
torch.cuda.empty_cache()
output_files = {}
for f in input_files:
audio_file_name = os.path.splitext(os.path.basename(f))[0]
output_files["vocals"] = os.path.join(options['output_folder'], audio_file_name + "_vocals.wav")
output_files["instrumental"] = os.path.join(options['output_folder'], audio_file_name + "_instrum.wav")
output_files["instrumental2"] = os.path.join(options['output_folder'], audio_file_name + "_instrum2.wav") # 第二个伴奏输出
output_files["bass"] = os.path.join(options['output_folder'], audio_file_name + "_bass.wav")
output_files["drums"] = os.path.join(options['output_folder'], audio_file_name + "_drums.wav")
output_files["other"] = os.path.join(options['output_folder'], audio_file_name + "_other.wav")
# 检查文件是否准备好
output_files_ready = []
for k, v in output_files.items():
if os.path.exists(v) and check_file_readiness(v):
output_files_ready.append(v)
else:
empty_data = np.zeros((44100, 2)) # 2个声道,1秒的静音,采样率为44100Hz
empty_file = tempfile.mktemp('.wav')
wavfile.write(empty_file, 44100, empty_data.astype(np.int16)) # 转换为int16,因为wavfile不支持float32
output_files_ready.append(empty_file)
print(len(output_files_ready)) # 应输出6
print("返回前")
return tuple(output_files_ready)
print("返回后")
description = """
# ZFTurbo Web-UI
Web-UI by [Ma5onic](https://github.com/Ma5onic)
## 选项:
- **仅使用CPU:** 如果GPU内存不足,请选择此项。速度会较慢。
- **使用单一ONNX模型:** 选择此项以使用单一ONNX模型。会稍微降低质量,但可以帮助减少GPU内存使用。
- **大块重叠:** 大块的重叠。根据需要调整。
- **小块重叠:** 小块的重叠。根据需要调整。
- **块大小:** 每次处理的块大小。如果遇到内存问题,请减少此值。
- **使用快速大GPU版本:** 选择此项以使用旧的快速方法,需要超过11GB的GPU内存。运行速度更快。
"""
theme = gr.themes.Base(
primary_hue="cyan",
secondary_hue="cyan",
)
with gr.Blocks(theme=theme) as demo:
gr.Markdown(description)
uploaded_file = gr.File(label="上传音频文件", type="file")
use_cpu = gr.Checkbox(label="仅使用CPU", value=True)
use_single_onnx = gr.Checkbox(label="使用单一ONNX模型", value=False)
large_overlap = gr.Number(label="大块重叠", value=0.6)
small_overlap = gr.Number(label="小块重叠", value=0.5)
chunk_size = gr.Number(label="块大小", value=1000000)
use_large_gpu = gr.Checkbox(label="使用快速大GPU版本", value=False)
process_button = gr.Button("处理音频")
vocals = gr.Audio(label="人声")
instrumental = gr.Audio(label="伴奏")
instrumental2 = gr.Audio(label="伴奏2")
bass = gr.Audio(label="贝斯")
drums = gr.Audio(label="鼓声")
other = gr.Audio(label="其他")
process_button.click(
separate_music_file_wrapper,
inputs=[uploaded_file, use_cpu, use_single_onnx, large_overlap, small_overlap, chunk_size, use_large_gpu],
outputs=[vocals, instrumental, instrumental2, bass, drums, other],
)
demo.queue().launch(debug=True, share=False)
|