Chitsanfei
commited on
Commit
·
3730a45
1
Parent(s):
f44f922
docs: blank
Browse files
app.py
CHANGED
@@ -47,11 +47,8 @@ with app:
|
|
47 |
with gr.TabItem("Basic"):
|
48 |
gr.Markdown(value="""
|
49 |
# sovits-emu-voice-transform | 可以变成凤笑梦的在线变声器
|
50 |
-
|
51 |
[](https://visitorbadge.io/status?path=https%3A%2F%2Fhuggingface.co%2Fspaces%2FMashiroSA%2Fsovits-emu-voice-transform)
|
52 |
-
|
53 |
-
---
|
54 |
-
|
55 |
**说明 / Introduction**
|
56 |
- 基于so-vits-svc 4.0的官方库示例修改而成。
|
57 |
- 所使用的音声训练集基于对话而来,因而转换后的音声在对话表现中会比乐曲中的人声中要好。
|
@@ -59,23 +56,18 @@ with app:
|
|
59 |
- Modified from the official library example based on so-vits-svc 4.0.
|
60 |
- The sound training set used is based on dialogue, thus the converted sound will perform better in dialogue than the vocals in the music.
|
61 |
- This project is conducted in no-profit.
|
62 |
-
|
63 |
```text
|
64 |
For academic purpose only and not for illegal purposes. We have no relationship or interest with SEGA or related organizations.
|
65 |
The model derivation output is only similar to Otori Emu and there is inevitable loss, which cannot be fully simulated.
|
66 |
If you have any questions, please send an email or forum for inquiry.
|
67 |
```
|
68 |
-
|
69 |
-
---
|
70 |
-
|
71 |
**如何使用**
|
72 |
- 如果用于日常说话时的对话转换,请提前录制一段低于90s的人声干声,上传,勾选下面的自动f0预测,其它的可以不用动,直接转换,过一会儿就能听到转换的声音了。
|
73 |
- 如果是乐曲中的人声,你可以使用自己的清唱,或者使用UVR5软件进行干声提取,上传,不要勾选自动f0预测,按情况进行变调(模型实际测试高于标准音C4的类似度较高,输入的干声是男声请+12,女声可以先不变),然后转换。
|
74 |
- 转换后的进度条右侧有个省略的点,在那边可以下载。
|
75 |
- 本repo的管理者 @MashiroSA 看不到你输入和输出后的内容,只有Hugging Face官方也许可以看到,请放心。
|
76 |
-
|
77 |
-
---
|
78 |
-
|
79 |
""")
|
80 |
spks = list(model.spk2id.keys())
|
81 |
sid = gr.Dropdown(label="音色", choices=spks, value=spks[0])
|
|
|
47 |
with gr.TabItem("Basic"):
|
48 |
gr.Markdown(value="""
|
49 |
# sovits-emu-voice-transform | 可以变成凤笑梦的在线变声器
|
|
|
50 |
[](https://visitorbadge.io/status?path=https%3A%2F%2Fhuggingface.co%2Fspaces%2FMashiroSA%2Fsovits-emu-voice-transform)
|
51 |
+
|
|
|
|
|
52 |
**说明 / Introduction**
|
53 |
- 基于so-vits-svc 4.0的官方库示例修改而成。
|
54 |
- 所使用的音声训练集基于对话而来,因而转换后的音声在对话表现中会比乐曲中的人声中要好。
|
|
|
56 |
- Modified from the official library example based on so-vits-svc 4.0.
|
57 |
- The sound training set used is based on dialogue, thus the converted sound will perform better in dialogue than the vocals in the music.
|
58 |
- This project is conducted in no-profit.
|
|
|
59 |
```text
|
60 |
For academic purpose only and not for illegal purposes. We have no relationship or interest with SEGA or related organizations.
|
61 |
The model derivation output is only similar to Otori Emu and there is inevitable loss, which cannot be fully simulated.
|
62 |
If you have any questions, please send an email or forum for inquiry.
|
63 |
```
|
64 |
+
|
|
|
|
|
65 |
**如何使用**
|
66 |
- 如果用于日常说话时的对话转换,请提前录制一段低于90s的人声干声,上传,勾选下面的自动f0预测,其它的可以不用动,直接转换,过一会儿就能听到转换的声音了。
|
67 |
- 如果是乐曲中的人声,你可以使用自己的清唱,或者使用UVR5软件进行干声提取,上传,不要勾选自动f0预测,按情况进行变调(模型实际测试高于标准音C4的类似度较高,输入的干声是男声请+12,女声可以先不变),然后转换。
|
68 |
- 转换后的进度条右侧有个省略的点,在那边可以下载。
|
69 |
- 本repo的管理者 @MashiroSA 看不到你输入和输出后的内容,只有Hugging Face官方也许可以看到,请放心。
|
70 |
+
- 关于下面选项中的聚类模型的使用:默认为0,值是0-1,越高越能贴近模型音色,但会导致咬字不清。
|
|
|
|
|
71 |
""")
|
72 |
spks = list(model.spk2id.keys())
|
73 |
sid = gr.Dropdown(label="音色", choices=spks, value=spks[0])
|