danieloneill commited on
Commit
7b1bc5d
·
1 Parent(s): bb9f357

Update README.md

Browse files

Formatting, wording, and happiness added

Files changed (1) hide show
  1. README.md +14 -8
README.md CHANGED
@@ -8,27 +8,33 @@ tags:
8
  - ddsp-svc
9
  ---
10
 
 
 
11
  These are a few test models I made using (and for use with) [DDSP-SVC](https://github.com/yxlllc/DDSP-SVC).
12
 
13
- I am not experienced with this software or technology,
14
 
15
- All examples are based on samples from an English speaker, though thanks to [DDSP](https://magenta.tensorflow.org/ddsp), they're generally fairly decent with use in a variety of other languages.
16
 
17
- All models are sampled at 44.1khz
18
 
19
- To use these, place the model file (model_XXXXXX.pt) and configuration file (config.yaml) in a directory.
20
 
21
- **It's rather important to mention that each model file should be in a distinct directory with its accompanying config.yaml or your results may be off/weird/broken.**
22
 
23
- Models:
24
  - PrimReaper - (Stereo) Trained on YouTube content from popular YouTuber "The Prim Reaper"
25
  - Panam - (Mono) Trained on extracted audio content from the Cyberpunk 2077 character dialogue named "Panam"
26
  - V-F - (Mono) Trained on extracted dialogue audio from the Female "V" character in Cyberpunk 2077
27
  - Nora - (Mono) Trained on Fallout 4 dialogue audio from the game character "Nora"
28
 
29
- If using DDSP-SVC's **gui_diff.py**, keep in mind that pitch adjustment is probably required if your voice is deeper than the character.
 
 
 
 
30
 
31
- Training is done following the suggestions and best practices according to the DDSP-SVC project, and K-steps are between 100 and 200.
32
 
33
  For realtime inference, my settings are generally as follows:
34
 
 
8
  - ddsp-svc
9
  ---
10
 
11
+ # Howdy
12
+
13
  These are a few test models I made using (and for use with) [DDSP-SVC](https://github.com/yxlllc/DDSP-SVC).
14
 
15
+ I am not experienced with this software or technology, but hope to provide samples which facilitate adoption and interest in this project and associated technologies.
16
 
17
+ All models are based on 44.1khz samples from a English speakers, though thanks to [DDSP](https://magenta.tensorflow.org/ddsp), they're generally fairly decent with use in a variety of other languages.
18
 
19
+ Training is done following the suggestions and best practices according to the DDSP-SVC project, with initial learning rates ranging between 0.00010 and 0.00020.
20
 
21
+ If using DDSP-SVC's **gui_diff.py**, keep in mind that pitch adjustment is probably required if your voice is deeper than the character.
22
 
23
+ For any/all questions/comments/suggestions, please use the Community section here.
24
 
25
+ ## Models
26
  - PrimReaper - (Stereo) Trained on YouTube content from popular YouTuber "The Prim Reaper"
27
  - Panam - (Mono) Trained on extracted audio content from the Cyberpunk 2077 character dialogue named "Panam"
28
  - V-F - (Mono) Trained on extracted dialogue audio from the Female "V" character in Cyberpunk 2077
29
  - Nora - (Mono) Trained on Fallout 4 dialogue audio from the game character "Nora"
30
 
31
+ ## Usage
32
+
33
+ To use these, place the model file (model_XXXXXX.pt) and configuration file (config.yaml) in a directory.
34
+
35
+ **It's rather important to mention that each model file should be in a distinct directory with its accompanying config.yaml or your results may be off/weird/broken.**
36
 
37
+ ## Settings
38
 
39
  For realtime inference, my settings are generally as follows:
40