Spaces:
Build error
Build error
Update assets/languages/en-US.json
Browse files- assets/languages/en-US.json +709 -709
assets/languages/en-US.json
CHANGED
@@ -1,710 +1,710 @@
|
|
1 |
-
{
|
2 |
-
"set_precision": "Override preprocessing and config.json to use {precision}.",
|
3 |
-
"set_lang": "Display language set to {lang}.",
|
4 |
-
"no_support_gpu": "Unfortunately, no compatible GPU is available to support your training.",
|
5 |
-
"text": "text",
|
6 |
-
"upload_success": "File {name} uploaded successfully.",
|
7 |
-
"download_url": "Download from the link",
|
8 |
-
"download_from_csv": "Download from the CSV model repository",
|
9 |
-
"download_from_applio": "Download model from Applio",
|
10 |
-
"upload": "Upload",
|
11 |
-
"option_not_valid": "Invalid option!",
|
12 |
-
"list_model": "Model list",
|
13 |
-
"success": "Completed!",
|
14 |
-
"index": "index",
|
15 |
-
"model": "model",
|
16 |
-
"zip": "compress",
|
17 |
-
"search": "search",
|
18 |
-
"provide_file": "Please provide a valid {filename} file!",
|
19 |
-
"start": "Starting {start}...",
|
20 |
-
"not_found": "Not found {name}.",
|
21 |
-
"found": "Found {results} results!",
|
22 |
-
"download_music": "download music",
|
23 |
-
"download": "download",
|
24 |
-
"provide_url": "Please provide a link.",
|
25 |
-
"provide_name_is_save": "Please provide a model name to save.",
|
26 |
-
"not_support_url": "Your model link is not supported.",
|
27 |
-
"error_occurred": "An error occurred: {e}.",
|
28 |
-
"not_model": "The file you uploaded is not a model file!",
|
29 |
-
"unable_analyze_model": "Unable to analyze the model!",
|
30 |
-
"credits": "\n**This project is brewed by [Phạm Huỳnh Anh]({author})**\n\n**The project is built based on several major projects such as:**\n\n**Conversion, Processing, Extraction, Training, Model Reading, Model Fusion, Training Model, Model Repository...: [Applio]({applio}) by the [AI Hispano]({ai_hispano}) group**\n\n**Extraction Methods, Information Display, Logging, Training Models...: [Retrieval-based-Voice-Conversion-WebUI]({rvc_webui}) by author [RVC BOSS]({rvc_boss})**\n\n**Music Separation Models MDX-Net and Demucs: [Python-audio-separator]({python_audio_separator}) by author [Andrew Beveridge]({andrew_beveridge})**\n",
|
31 |
-
"download_pretrain": "Downloading pre-trained model...",
|
32 |
-
"provide_pretrain": "Please provide a pre-trained model link {dg}.",
|
33 |
-
"provide_hubert": "Please provide a link to the embedding model.",
|
34 |
-
"sr_not_same": "The sample rates of the two models are not the same.",
|
35 |
-
"architectures_not_same": "Cannot merge models. The architectures are not the same.",
|
36 |
-
"fushion_model": "model fusion",
|
37 |
-
"model_fushion_info": "The model {name} is fused from {pth_1} and {pth_2} with a ratio of {ratio}.",
|
38 |
-
"not_found_create_time": "Creation time not found.",
|
39 |
-
"format_not_valid": "Invalid format.",
|
40 |
-
"read_info": "Models trained on different applications may produce different information or may not be readable!",
|
41 |
-
"epoch": "epoch.",
|
42 |
-
"step": "step",
|
43 |
-
"sr": "sample rate.",
|
44 |
-
"f0": "pitch training",
|
45 |
-
"version": "version.",
|
46 |
-
"not_f0": "Pitch training not performed",
|
47 |
-
"trained_f0": "Pitch training performed",
|
48 |
-
"model_info": "Model Name: {model_name}\n\n Model Creator: {model_author}\n\nEpoch: {epochs}\n\nSteps: {steps}\n\nVersion: {version}\n\nSample Rate: {sr}\n\nPitch Training: {pitch_guidance}\n\nHash (ID): {model_hash}\n\nCreation Time: {creation_date_str}\n",
|
49 |
-
"input_not_valid": "Please provide valid input!",
|
50 |
-
"output_not_valid": "Please provide valid output!",
|
51 |
-
"apply_effect": "apply effect",
|
52 |
-
"enter_the_text": "Please enter the text to read!",
|
53 |
-
"choose_voice": "Please choose a voice!",
|
54 |
-
"convert": "Converting {name}...",
|
55 |
-
"separator_music": "music separation",
|
56 |
-
"notfound": "Not found",
|
57 |
-
"turn_on_use_audio": "Please enable using separated audio to proceed",
|
58 |
-
"turn_off_convert_backup": "Disable backup voice conversion to use the original voice",
|
59 |
-
"turn_off_merge_backup": "Disable merging backup voice to use the original voice",
|
60 |
-
"not_found_original_vocal": "Original vocal not found!",
|
61 |
-
"convert_vocal": "Converting voice...",
|
62 |
-
"convert_success": "Voice conversion completed!",
|
63 |
-
"convert_backup": "Converting backup voice...",
|
64 |
-
"convert_backup_success": "Backup voice conversion completed!",
|
65 |
-
"merge_backup": "Merging main voice with backup voice...",
|
66 |
-
"merge_success": "Merge completed.",
|
67 |
-
"is_folder": "Input is a folder: Converting all audio files in the folder...",
|
68 |
-
"not_found_in_folder": "No audio files found in the folder!",
|
69 |
-
"batch_convert": "Batch conversion in progress...",
|
70 |
-
"batch_convert_success": "Batch conversion successful!",
|
71 |
-
"create": "create",
|
72 |
-
"provide_name": "Please provide a model name.",
|
73 |
-
"not_found_data": "Data not found",
|
74 |
-
"not_found_data_preprocess": "Processed audio data not found, please reprocess.",
|
75 |
-
"not_found_data_extract": "Extracted audio data not found, please re-extract.",
|
76 |
-
"provide_pretrained": "Please provide pre-trained {dg}.",
|
77 |
-
"download_pretrained": "Download pre-trained {dg}{rvc_version} original",
|
78 |
-
"not_found_pretrain": "Pre-trained {dg} not found",
|
79 |
-
"not_use_pretrain": "No pre-trained model will be used",
|
80 |
-
"training": "training",
|
81 |
-
"display_title": "<h1> 🎵 Voice model conversion and training interface created by Anh 🎵 <h1>",
|
82 |
-
"rick_roll": "Click here if you want to be Rick
|
83 |
-
"terms_of_use": "**Please do not use the project for any unethical, illegal, or harmful purposes to individuals or organizations...**",
|
84 |
-
"exemption": "**In cases where users do not comply with the terms or violate them, I will not be responsible for any claims, damages, or liabilities, whether in contract, negligence, or other causes arising from, outside of, or related to the software, its use, or other transactions associated with it.**",
|
85 |
-
"separator_tab": "Music Separation",
|
86 |
-
"4_part": "A simple music separation system can separate into 4 parts: Music, vocals, main vocals, backup vocals",
|
87 |
-
"clear_audio": "Clean audio",
|
88 |
-
"separator_backing": "Separate backup vocals",
|
89 |
-
"denoise_backing": "Denoise backup vocals",
|
90 |
-
"denoise_mdx": "Denoise MDX separation",
|
91 |
-
"use_mdx": "Use MDX",
|
92 |
-
"dereveb_audio": "Remove vocal reverb",
|
93 |
-
"dereveb_backing": "Remove backup reverb",
|
94 |
-
"denoise_dereveb": "Denoise reverb",
|
95 |
-
"separator_model": "Music separation model",
|
96 |
-
"separator_backing_model": "Backup separation model",
|
97 |
-
"shift": "Number of predictions",
|
98 |
-
"shift_info": "Higher is better quality but slower and uses more resources",
|
99 |
-
"segments_size": "Segment Size",
|
100 |
-
"segments_size_info": "Higher is better quality but uses more resources",
|
101 |
-
"batch_size": "Batch size",
|
102 |
-
"batch_size_info": "Number of samples processed simultaneously in one training cycle. Higher can cause memory overflow",
|
103 |
-
"mdx_batch_size_info": "Number of samples processed at a time. Batch processing optimizes calculations. Large batches can cause memory overflow; small batches reduce resource efficiency",
|
104 |
-
"overlap": "Overlap",
|
105 |
-
"overlap_info": "Overlap amount between prediction windows",
|
106 |
-
"export_format": "Audio format",
|
107 |
-
"export_info": "The audio format to export the audio file in",
|
108 |
-
"output_separator": "Separated audio",
|
109 |
-
"hop_length_info": "Indicates the time interval for shifting the analysis window on the audio signal when performing transformations. Smaller values increase detail but require more computational resources",
|
110 |
-
"drop_audio": "Drop audio here",
|
111 |
-
"drop_text": "Drop text file here",
|
112 |
-
"use_url": "Use YouTube link",
|
113 |
-
"url_audio": "Link to audio",
|
114 |
-
"downloads": "Downloads",
|
115 |
-
"clean_strength": "Audio cleaning strength",
|
116 |
-
"clean_strength_info": "Strength of the audio cleaner for filtering vocals during export",
|
117 |
-
"input_output": "Audio input, output",
|
118 |
-
"audio_path": "Input audio path",
|
119 |
-
"refesh": "Refresh",
|
120 |
-
"output_folder": "Output audio folder path",
|
121 |
-
"output_folder_info": "Enter the folder path where the audio will be exported",
|
122 |
-
"input_audio": "Audio input",
|
123 |
-
"instruments": "Instruments",
|
124 |
-
"original_vocal": "Original vocal",
|
125 |
-
"main_vocal": "Main vocal",
|
126 |
-
"backing_vocal": "Backup vocal",
|
127 |
-
"convert_audio": "Convert Audio",
|
128 |
-
"convert_info": "Convert audio using a trained voice model",
|
129 |
-
"upscale_audio": "Upscale quality",
|
130 |
-
"autotune": "Auto-tune",
|
131 |
-
"use_audio": "Use separated audio",
|
132 |
-
"convert_original": "Convert original voice",
|
133 |
-
"convert_backing": "Convert backup voice",
|
134 |
-
"not_merge_backing": "Do not merge backup voice",
|
135 |
-
"merge_instruments": "Merge instruments",
|
136 |
-
"pitch": "Pitch",
|
137 |
-
"pitch_info": "Recommendation: set to 12 to change male voice to female and vice versa",
|
138 |
-
"model_accordion": "Model and index",
|
139 |
-
"model_name": "Model file",
|
140 |
-
"index_path": "Index file",
|
141 |
-
"index_strength": "Index strength",
|
142 |
-
"index_strength_info": "Higher values increase strength. However, lower values may reduce artificial effects in the audio",
|
143 |
-
"output_path": "Audio output path",
|
144 |
-
"output_path_info": "Enter the output path (leave it as .wav format; it will auto-correct during conversion)",
|
145 |
-
"setting": "General settings",
|
146 |
-
"f0_method": "Extraction method",
|
147 |
-
"f0_method_info": "Method used for data extraction",
|
148 |
-
"f0_method_hybrid": "HYBRID extraction method",
|
149 |
-
"f0_method_hybrid_info": "Combines two different extraction types",
|
150 |
-
"hubert_model": "Embedding model",
|
151 |
-
"hubert_info": "Pre-trained model to assist embedding",
|
152 |
-
"modelname": "Model name",
|
153 |
-
"modelname_info": "If you have your own model, just upload it and input the name here",
|
154 |
-
"split_audio": "Split audio",
|
155 |
-
"split_audio_info": "Splitting audio into smaller parts for conversion can speed up the process",
|
156 |
-
"batch_process": "Batch processing",
|
157 |
-
"batch_process_info": "Batch processing can reduce training time but may overload resources",
|
158 |
-
"autotune_rate": "Auto-tune level",
|
159 |
-
"autotune_rate_info": "Level of auto-tuning adjustment",
|
160 |
-
"resample": "Resample",
|
161 |
-
"resample_info": "Resample post-processing to the final sample rate; 0 means no resampling",
|
162 |
-
"filter_radius": "Median filter",
|
163 |
-
"filter_radius_info": "If greater than three, median filtering is applied. The value represents the filter radius and can reduce breathiness or noise.",
|
164 |
-
"volume_envelope": "Volume envelope",
|
165 |
-
"volume_envelope_info": "Use the input volume envelope to replace or mix with the output volume envelope. The closer to 1, the more the output envelope is used",
|
166 |
-
"protect": "Consonant protection",
|
167 |
-
"protect_info": "Protect distinct consonants and breathing sounds to prevent audio tearing and other artifacts. Increasing this value provides comprehensive protection. Reducing it may reduce protection but also minimize indexing effects",
|
168 |
-
"output_convert": "Converted audio",
|
169 |
-
"main_convert": "Convert main voice",
|
170 |
-
"main_or_backing": "Main voice + Backup voice",
|
171 |
-
"voice_or_instruments": "Voice + Instruments",
|
172 |
-
"convert_text": "Convert Text",
|
173 |
-
"convert_text_markdown": "## Convert Text to Speech",
|
174 |
-
"convert_text_markdown_2": "Convert text to speech and read aloud using the trained voice model",
|
175 |
-
"input_txt": "Input data from a text file (.txt)",
|
176 |
-
"text_to_speech": "Text to read",
|
177 |
-
"voice_speed": "Reading speed",
|
178 |
-
"voice_speed_info": "Speed of the voice",
|
179 |
-
"tts_1": "1. Convert Text",
|
180 |
-
"tts_2": "2. Convert Speech",
|
181 |
-
"voice": "Voices by country",
|
182 |
-
"output_tts": "Output speech path",
|
183 |
-
"output_tts_convert": "Converted speech output path",
|
184 |
-
"tts_output": "Enter the output path",
|
185 |
-
"output_tts_markdown": "Unconverted and converted audio",
|
186 |
-
"output_text_to_speech": "Generated speech from text-to-speech conversion",
|
187 |
-
"output_file_tts_convert": "Speech converted using the model",
|
188 |
-
"output_audio": "Audio output",
|
189 |
-
"provide_output": "Enter the output path",
|
190 |
-
"audio_effects": "Audio Effects",
|
191 |
-
"apply_audio_effects": "## Add Additional Audio Effects",
|
192 |
-
"audio_effects_edit": "Edit and add effects to audio",
|
193 |
-
"reverb": "Reverb effect",
|
194 |
-
"chorus": "Chorus effect",
|
195 |
-
"delay": "Delay effect",
|
196 |
-
"more_option": "Additional options",
|
197 |
-
"phaser": "Phaser effect",
|
198 |
-
"compressor": "Compressor effect",
|
199 |
-
"apply": "Apply",
|
200 |
-
"reverb_freeze": "Freeze mode",
|
201 |
-
"reverb_freeze_info": "Create a continuous echo effect when this mode is enabled",
|
202 |
-
"room_size": "Room size",
|
203 |
-
"room_size_info": "Adjust the room space to create reverberation",
|
204 |
-
"damping": "Damping",
|
205 |
-
"damping_info": "Adjust the level of absorption to control the amount of reverberation",
|
206 |
-
"wet_level": "Reverb signal level",
|
207 |
-
"wet_level_info": "Adjust the level of the reverb signal effect",
|
208 |
-
"dry_level": "Original signal level",
|
209 |
-
"dry_level_info": "Adjust the level of the signal without effects",
|
210 |
-
"width": "Audio width",
|
211 |
-
"width_info": "Adjust the width of the audio space",
|
212 |
-
"chorus_depth": "Chorus depth",
|
213 |
-
"chorus_depth_info": "Adjust the intensity of the chorus to create a wider sound",
|
214 |
-
"chorus_rate_hz": "Frequency",
|
215 |
-
"chorus_rate_hz_info": "Adjust the oscillation speed of the chorus effect",
|
216 |
-
"chorus_mix": "Mix signals",
|
217 |
-
"chorus_mix_info": "Adjust the mix level between the original and the processed signal",
|
218 |
-
"chorus_centre_delay_ms": "Center delay (ms)",
|
219 |
-
"chorus_centre_delay_ms_info": "The delay time between stereo channels to create the chorus effect",
|
220 |
-
"chorus_feedback": "Feedback",
|
221 |
-
"chorus_feedback_info": "Adjust the amount of the effect signal fed back into the original signal",
|
222 |
-
"delay_seconds": "Delay time",
|
223 |
-
"delay_seconds_info": "Adjust the delay time between the original and the processed signal",
|
224 |
-
"delay_feedback": "Delay feedback",
|
225 |
-
"delay_feedback_info": "Adjust the amount of feedback signal, creating a repeating effect",
|
226 |
-
"delay_mix": "Delay signal mix",
|
227 |
-
"delay_mix_info": "Adjust the mix level between the original and delayed signal",
|
228 |
-
"fade": "Fade effect",
|
229 |
-
"bass_or_treble": "Bass and treble",
|
230 |
-
"limiter": "Threshold limiter",
|
231 |
-
"distortion": "Distortion effect",
|
232 |
-
"gain": "Audio gain",
|
233 |
-
"bitcrush": "Bit reduction effect",
|
234 |
-
"clipping": "Clipping effect",
|
235 |
-
"fade_in": "Fade-in effect (ms)",
|
236 |
-
"fade_in_info": "Time for the audio to gradually increase from 0 to normal level",
|
237 |
-
"fade_out": "Fade-out effect (ms)",
|
238 |
-
"fade_out_info": "Time for the audio to gradually decrease to 0",
|
239 |
-
"bass_boost": "Bass boost level (dB)",
|
240 |
-
"bass_boost_info": "Boost level for bass frequencies",
|
241 |
-
"bass_frequency": "Low-pass filter cutoff frequency (Hz)",
|
242 |
-
"bass_frequency_info": "Frequency below which the sound will start to reduce. Lower frequencies make bass clearer",
|
243 |
-
"treble_boost": "Treble boost level (dB)",
|
244 |
-
"treble_boost_info": "Boost level for high frequencies",
|
245 |
-
"treble_frequency": "High-pass filter cutoff frequency (Hz)",
|
246 |
-
"treble_frequency_info": "Sounds below this frequency will be filtered out. Higher frequencies retain more treble",
|
247 |
-
"limiter_threashold_db": "Limiter threshold",
|
248 |
-
"limiter_threashold_db_info": "Limit the maximum audio level to prevent it from exceeding the threshold",
|
249 |
-
"limiter_release_ms": "Release time",
|
250 |
-
"limiter_release_ms_info": "Time for the audio to return after being limited",
|
251 |
-
"distortion_info": "Adjust the level of distortion to create a noisy effect",
|
252 |
-
"gain_info": "Adjust the volume level of the signal",
|
253 |
-
"clipping_threashold_db": "Clipping threshold",
|
254 |
-
"clipping_threashold_db_info": "Trim signals exceeding the threshold, creating a distorted sound",
|
255 |
-
"bitcrush_bit_depth": "Bit depth",
|
256 |
-
"bitcrush_bit_depth_info": "Reduce audio quality by decreasing bit depth, creating a distorted effect",
|
257 |
-
"phaser_depth": "Phaser depth",
|
258 |
-
"phaser_depth_info": "Adjust the depth of the effect, impacting its intensity",
|
259 |
-
"phaser_rate_hz": "Frequency",
|
260 |
-
"phaser_rate_hz_info": "Adjust the frequency of the phaser effect",
|
261 |
-
"phaser_mix": "Mix signal",
|
262 |
-
"phaser_mix_info": "Adjust the mix level between the original and processed signals",
|
263 |
-
"phaser_centre_frequency_hz": "Center frequency",
|
264 |
-
"phaser_centre_frequency_hz_info": "The center frequency of the phaser effect, affecting the adjusted frequencies",
|
265 |
-
"phaser_feedback": "Feedback",
|
266 |
-
"phaser_feedback_info": "Adjust the feedback level of the effect, creating a stronger or lighter phaser feel",
|
267 |
-
"compressor_threashold_db": "Compressor threshold",
|
268 |
-
"compressor_threashold_db_info": "The threshold level above which the audio will be compressed",
|
269 |
-
"compressor_ratio": "Compression ratio",
|
270 |
-
"compressor_ratio_info": "Adjust the level of audio compression when exceeding the threshold",
|
271 |
-
"compressor_attack_ms": "Attack time (ms)",
|
272 |
-
"compressor_attack_ms_info": "Time for compression to start taking effect after the audio exceeds the threshold",
|
273 |
-
"compressor_release_ms": "Release time",
|
274 |
-
"compressor_release_ms_info": "Time for the audio to return to normal after being compressed",
|
275 |
-
"create_dataset_url": "Link to audio (use commas for multiple links)",
|
276 |
-
"createdataset": "Create dataset",
|
277 |
-
"create_dataset_markdown": "## Create Training Data from YouTube",
|
278 |
-
"create_dataset_markdown_2": "Process and create training datasets using YouTube links",
|
279 |
-
"denoise": "Model denoise",
|
280 |
-
"skip": "Skip seconds",
|
281 |
-
"model_ver": "Voice separation version",
|
282 |
-
"model_ver_info": "The model version for separating vocals",
|
283 |
-
"create_dataset_info": "Dataset creation information",
|
284 |
-
"output_data": "Dataset output",
|
285 |
-
"output_data_info": "Output data after creation",
|
286 |
-
"skip_start": "Skip beginning",
|
287 |
-
"skip_start_info": "Skip the initial seconds of the audio; use commas for multiple audios",
|
288 |
-
"skip_end": "Skip end",
|
289 |
-
"skip_end_info": "Skip the final seconds of the audio; use commas for multiple audios",
|
290 |
-
"training_model": "Train Model",
|
291 |
-
"training_markdown": "Train and build a voice model with a set of voice data",
|
292 |
-
"training_model_name": "Name of the model during training (avoid special characters or spaces)",
|
293 |
-
"sample_rate": "Sample rate",
|
294 |
-
"sample_rate_info": "Sample rate of the model",
|
295 |
-
"training_version": "Model version",
|
296 |
-
"training_version_info": "Version of the model during training",
|
297 |
-
"training_pitch": "Pitch training",
|
298 |
-
"training_pitch_info": "Train the pitch for the model",
|
299 |
-
"upload_dataset": "Upload training data",
|
300 |
-
"preprocess_split": "Should be disabled if data has already been processed",
|
301 |
-
"preprocess_effect": "Processing effects",
|
302 |
-
"preprocess_effect_info": "Should be disabled if data has already been processed",
|
303 |
-
"clear_dataset": "Clean dataset",
|
304 |
-
"clear_dataset_info": "Clean data segments",
|
305 |
-
"preprocess_info": "Preprocessing information",
|
306 |
-
"preprocess_button": "1. Process data",
|
307 |
-
"extract_button": "2. Extract data",
|
308 |
-
"extract_info": "Data extraction information",
|
309 |
-
"total_epoch": "Total epochs",
|
310 |
-
"total_epoch_info": "Total training epochs",
|
311 |
-
"save_epoch": "Save frequency",
|
312 |
-
"save_epoch_info": "Frequency of saving the model during training to allow retraining",
|
313 |
-
"create_index": "Create index",
|
314 |
-
"index_algorithm": "Index algorithm",
|
315 |
-
"index_algorithm_info": "Algorithm for creating the index",
|
316 |
-
"custom_dataset": "Custom folder",
|
317 |
-
"custom_dataset_info": "Custom folder for training data",
|
318 |
-
"overtraining_detector": "Overtraining detector",
|
319 |
-
"overtraining_detector_info": "Check for overtraining during model training",
|
320 |
-
"sync_graph": "Synchronize graph",
|
321 |
-
"sync_graph_info": "Only enable if you need to retrain the model from scratch.",
|
322 |
-
"cache_in_gpu": "Cache model in GPU",
|
323 |
-
"cache_in_gpu_info": "Store the model in GPU cache memory",
|
324 |
-
"dataset_folder": "Folder containing data",
|
325 |
-
"threshold": "Overtraining threshold",
|
326 |
-
"setting_cpu_gpu": "CPU/GPU settings",
|
327 |
-
"gpu_number": "Number of GPUs used",
|
328 |
-
"gpu_number_info": "Number of GPUs used during training",
|
329 |
-
"save_only_latest": "Save only the latest",
|
330 |
-
"save_only_latest_info": "Save only the latest D and G models",
|
331 |
-
"save_every_weights": "Save all models",
|
332 |
-
"save_every_weights_info": "Save all models after each epoch",
|
333 |
-
"gpu_info": "GPU information",
|
334 |
-
"gpu_info_2": "Information about the GPU used during training",
|
335 |
-
"cpu_core": "Number of CPU cores available",
|
336 |
-
"cpu_core_info": "Number of CPU cores used during training",
|
337 |
-
"not_use_pretrain_2": "Do not use pretraining",
|
338 |
-
"not_use_pretrain_info": "Do not use pre-trained models",
|
339 |
-
"custom_pretrain": "Custom pretraining",
|
340 |
-
"custom_pretrain_info": "Customize pre-training settings",
|
341 |
-
"pretrain_file": "Pre-trained model file {dg}",
|
342 |
-
"train_info": "Training information",
|
343 |
-
"export_model": "5. Export Model",
|
344 |
-
"zip_model": "2. Compress model",
|
345 |
-
"output_zip": "Output file after compression",
|
346 |
-
"refesh_pretrain": "Refresh pre-trained model",
|
347 |
-
"model_path": "Model path",
|
348 |
-
"model_ratio": "Model ratio",
|
349 |
-
"model_ratio_info": "Adjusting towards one side will make the model more like that side",
|
350 |
-
"output_model_path": "Model output path",
|
351 |
-
"fushion": "Model Fusion",
|
352 |
-
"fushion_markdown": "## Merge Two Models",
|
353 |
-
"fushion_markdown_2": "Combine two voice models into a single model",
|
354 |
-
"read_model": "Read Information",
|
355 |
-
"read_model_markdown": "## Read Model Information",
|
356 |
-
"read_model_markdown_2": "Retrieve recorded information within the model",
|
357 |
-
"drop_model": "Drop model here",
|
358 |
-
"readmodel": "Read model",
|
359 |
-
"model_path_info": "Enter the path to the model file",
|
360 |
-
"modelinfo": "Model Information",
|
361 |
-
"download_markdown": "## Download Model",
|
362 |
-
"download_markdown_2": "Download voice models, pre-trained models, and embedding models",
|
363 |
-
"model_download": "Download voice model",
|
364 |
-
"model_url": "Link to the model",
|
365 |
-
"30s": "Please wait about 30 seconds. The system will restart automatically!",
|
366 |
-
"fp_select": "Setting precision...",
|
367 |
-
"fp_select_2": "Precision set to: {fp}",
|
368 |
-
"model_download_select": "Choose a model download method",
|
369 |
-
"model_warehouse": "Model repository",
|
370 |
-
"get_model": "Retrieve model",
|
371 |
-
"name_to_search": "Name to search",
|
372 |
-
"search_2": "Search",
|
373 |
-
"select_download_model": "Choose a searched model (Click to select)",
|
374 |
-
"download_pretrainec": "Download pre-trained model",
|
375 |
-
"only_huggingface": "Supports only huggingface.co",
|
376 |
-
"pretrained_url": "Pre-trained model link {dg}",
|
377 |
-
"select_pretrain": "Choose pre-trained model",
|
378 |
-
"select_pretrain_info": "Choose a pre-trained model to download",
|
379 |
-
"pretrain_sr": "Model sample rate",
|
380 |
-
"drop_pretrain": "Drop pre-trained model {dg} here",
|
381 |
-
"hubert_download": "Download embedding model",
|
382 |
-
"hubert_url": "Link to embedding model",
|
383 |
-
"drop_hubert": "Drop embedding model here",
|
384 |
-
"source": "Origin",
|
385 |
-
"source_info": "## Project Origin and Author",
|
386 |
-
"settings": "Settings",
|
387 |
-
"settings_markdown": "## Additional Settings",
|
388 |
-
"settings_markdown_2": "Customize additional features of the project",
|
389 |
-
"lang": "Language",
|
390 |
-
"lang_restart": "The display language in the project (When changing the language, the system will automatically restart after 30 seconds to update)",
|
391 |
-
"change_lang": "Change Language",
|
392 |
-
"fp_train": "Training precision",
|
393 |
-
"fp_info": "Use fp32 for higher precision but requires more memory and strong computation. Use fp16 to reduce memory usage and increase processing speed.",
|
394 |
-
"fp_button": "Change Precision",
|
395 |
-
"theme": "Theme",
|
396 |
-
"theme_restart": "Theme type displayed in the interface (When changing the theme, the system will automatically restart after 30 seconds to update)",
|
397 |
-
"theme_button": "Change Theme",
|
398 |
-
"change_light_dark": "Switch Light/Dark Mode",
|
399 |
-
"tensorboard_url": "Graph URL",
|
400 |
-
"errors_loading_audio": "Error loading audio: {e}",
|
401 |
-
"apply_error": "An error occurred while applying effects: {e}",
|
402 |
-
"indexpath": "Index path",
|
403 |
-
"skip_file": "Part {i} skipped because it is too short: {chunk}ms",
|
404 |
-
"split_total": "Total parts split",
|
405 |
-
"process_audio_error": "An error occurred while processing the audio",
|
406 |
-
"merge_error": "An error occurred while merging audio",
|
407 |
-
"not_found_convert_file": "Processed file not found",
|
408 |
-
"convert_batch": "Batch conversion...",
|
409 |
-
"found_audio": "Found {audio_files} audio files for conversion.",
|
410 |
-
"not_found_audio": "No audio files found!",
|
411 |
-
"error_convert": "An error occurred during audio conversion: {e}",
|
412 |
-
"error_convert_batch": "An error occurred during the conversion of audio segments: {e}",
|
413 |
-
"error_convert_batch_2": "An error occurred during batch audio conversion: {e}",
|
414 |
-
"convert_batch_success": "Batch conversion completed successfully in {elapsed_time} seconds. {output_path}",
|
415 |
-
"convert_audio_success": "File {input_path} converted successfully in {elapsed_time} seconds. {output_path}",
|
416 |
-
"hybrid_methods": "Estimating f0 pitch using methods {methods}",
|
417 |
-
"method_not_valid": "Invalid method",
|
418 |
-
"read_faiss_index_error": "An error occurred while reading the FAISS index: {e}",
|
419 |
-
"read_model_error": "Failed to load model: {e}",
|
420 |
-
"denoise_error": "An error occurred while denoising: {e}",
|
421 |
-
"change_format": "Converting audio to format {output_format}...",
|
422 |
-
"change_format_error": "An error occurred while converting audio format: {e}",
|
423 |
-
"starting_download": "Starting download",
|
424 |
-
"version_not_valid": "Invalid vocal separation version",
|
425 |
-
"create_dataset_value_not_valid": "Enable music separation option to use reverb removal options",
|
426 |
-
"skip<audio": "Cannot skip as skip time is less than audio file length",
|
427 |
-
"skip>audio": "Cannot skip as skip time is greater than audio file length",
|
428 |
-
"=<0": "Skip time is less than or equal to 0 and has been skipped",
|
429 |
-
"skip_warning": "Skip duration ({seconds} seconds) exceeds audio length ({total_duration} seconds). Skipping.",
|
430 |
-
"download_success": "Download completed successfully",
|
431 |
-
"create_dataset_error": "An error occurred while creating the training dataset",
|
432 |
-
"create_dataset_success": "Training dataset creation completed in {elapsed_time} seconds",
|
433 |
-
"skip_start_audio": "Successfully skipped start of audio: {input_file}",
|
434 |
-
"skip_end_audio": "Successfully skipped end of audio: {input_file}",
|
435 |
-
"merge_audio": "Merged all parts containing audio",
|
436 |
-
"separator_process": "Separating vocals: {input}...",
|
437 |
-
"not_found_main_vocal": "Main vocal not found!",
|
438 |
-
"not_found_backing_vocal": "Backup vocal not found!",
|
439 |
-
"not_found_instruments": "Instruments not found",
|
440 |
-
"merge_instruments_process": "Merging vocals with instruments...",
|
441 |
-
"dereverb": "Removing vocal reverb",
|
442 |
-
"dereverb_success": "Successfully removed vocal reverb",
|
443 |
-
"save_index": "Index file saved",
|
444 |
-
"create_index_error": "An error occurred while creating the index",
|
445 |
-
"sr_not_16000": "Sample rate must be 16000",
|
446 |
-
"gpu_not_valid": "Invalid GPU index. Switching to CPU.",
|
447 |
-
"extract_file_error": "An error occurred while extracting the file",
|
448 |
-
"extract_f0_method": "Starting pitch extraction using {num_processes} cores with method {f0_method}...",
|
449 |
-
"extract_f0": "Pitch Extraction",
|
450 |
-
"extract_f0_success": "Pitch extraction completed in {elapsed_time} seconds.",
|
451 |
-
"NaN": "contains NaN values and will be ignored.",
|
452 |
-
"start_extract_hubert": "Starting Hubert embedding extraction...",
|
453 |
-
"not_found_audio_file": "Audio file not found. Please ensure you provided the correct audio.",
|
454 |
-
"extract_hubert": "Extract Hubert embeddings",
|
455 |
-
"process_error": "An error occurred during processing",
|
456 |
-
"extract_hubert_success": "Hubert embedding extraction completed in {elapsed_time} seconds.",
|
457 |
-
"export_process": "Model path",
|
458 |
-
"extract_error": "An error occurred during data extraction",
|
459 |
-
"extract_success": "Data extraction successful",
|
460 |
-
"min_length>=min_interval>=hop_size": "min_length must be greater than or equal to min_interval and hop_size",
|
461 |
-
"max_sil_kept>=hop_size": "max_sil_kept must be greater than or equal to hop_size",
|
462 |
-
"start_preprocess": "Starting data preprocessing with {num_processes} cores...",
|
463 |
-
"not_integer": "Voice ID folder must be an integer; instead got",
|
464 |
-
"preprocess": "Preprocessing Data",
|
465 |
-
"preprocess_success": "Preprocessing completed in {elapsed_time} seconds.",
|
466 |
-
"preprocess_model_success": "Preprocessing data for the model completed successfully",
|
467 |
-
"turn_on_dereverb": "Reverb removal for backup vocals requires enabling reverb removal",
|
468 |
-
"turn_on_separator_backing": "Backup vocal separation requires enabling vocal separation",
|
469 |
-
"demucs_model": "Demucs model",
|
470 |
-
"backing_model_ver": "Backup vocal separation model version",
|
471 |
-
"mdx_model": "MDX model",
|
472 |
-
"clean_audio_success": "Audio cleaned successfully!",
|
473 |
-
"separator_error": "An error occurred during music separation",
|
474 |
-
"separator_success": "Music separation completed in {elapsed_time} seconds",
|
475 |
-
"separator_process_2": "Processing music separation",
|
476 |
-
"separator_success_2": "Music separation successful!",
|
477 |
-
"separator_process_backing": "Processing backup vocal separation",
|
478 |
-
"separator_process_backing_success": "Backup vocal separation successful!",
|
479 |
-
"process_original": "Processing original vocal reverb removal...",
|
480 |
-
"process_original_success": "Original vocal reverb removal successful!",
|
481 |
-
"process_main": "Processing main vocal reverb removal...",
|
482 |
-
"process_main_success": "Main vocal reverb removal successful!",
|
483 |
-
"process_backing": "Processing backup vocal reverb removal...",
|
484 |
-
"process_backing_success": "Backup vocal reverb removal successful!",
|
485 |
-
"save_every_epoch": "Save model after: {save_every_epoch} epochs",
|
486 |
-
"total_e": "Total training epochs: {total_epoch} epochs",
|
487 |
-
"dorg": "Pre-trained G: {pretrainG} | Pre-trained D: {pretrainD}",
|
488 |
-
"training_f0": "Pitch training",
|
489 |
-
"not_gpu": "No GPU detected, reverting to CPU (not recommended)",
|
490 |
-
"sync": "Graph synchronization is now enabled! When synchronization is active, the model will undergo a single training phase. After graphs are synchronized, training will resume with previously specified epochs.",
|
491 |
-
"sync_success": "Graph synchronization successful!",
|
492 |
-
"not_found_checkpoint": "Checkpoint file not found: {checkpoint_path}",
|
493 |
-
"save_checkpoint": "Reloaded checkpoint '{checkpoint_path}' (epoch {checkpoint_dict})",
|
494 |
-
"save_model": "Saved model '{checkpoint_path}' (epoch {iteration})",
|
495 |
-
"sid_error": "Error converting speaker ID '{sid}' to integer. Exception: {e}",
|
496 |
-
"sr_does_not_match": "{sample_rate} Sample rate does not match target {sample_rate2} Sample rate",
|
497 |
-
"spec_error": "An error occurred while retrieving specifications from {spec_filename}: {e}",
|
498 |
-
"time_or_speed_training": "time={current_time} | training speed={elapsed_time_str}",
|
499 |
-
"savemodel": "Saved model '{model_dir}' (epoch {epoch} and step {step})",
|
500 |
-
"model_author": "Credit model to {model_author}",
|
501 |
-
"unregistered": "Model unregistered",
|
502 |
-
"not_author": "Model not credited",
|
503 |
-
"training_author": "Model creator name",
|
504 |
-
"training_author_info": "To credit the model, enter your name here",
|
505 |
-
"extract_model_error": "An error occurred while extracting the model",
|
506 |
-
"start_training": "Starting training",
|
507 |
-
"import_pretrain": "Loaded pre-trained model ({dg}) '{pretrain}'",
|
508 |
-
"not_using_pretrain": "No pre-trained model ({dg}) will be used",
|
509 |
-
"training_warning": "WARNING: Generated loss is lower than the lower threshold loss for the next epoch.",
|
510 |
-
"overtraining_find": "Overtraining detected at epoch {epoch} with smoothed generator loss {smoothed_value_gen} and smoothed discriminator loss {smoothed_value_disc}",
|
511 |
-
"best_epoch": "New best epoch {epoch} with smoothed generator loss {smoothed_value_gen} and smoothed discriminator loss {smoothed_value_disc}",
|
512 |
-
"success_training": "Training completed with {epoch} epochs, {global_step} steps, and {loss_gen_all} total generator loss.",
|
513 |
-
"training_info": "Lowest generator loss: {lowest_value_rounded} at epoch {lowest_value_epoch}, step {lowest_value_step}",
|
514 |
-
"model_training_info": "{model_name} | epoch={epoch} | step={global_step} | {epoch_recorder} | lowest value={lowest_value_rounded} (epoch {lowest_value_epoch} and step {lowest_value_step}) | remaining epochs for overtraining: g/total: {remaining_epochs_gen} d/total: {remaining_epochs_disc} | smoothed generator loss={smoothed_value_gen} | smoothed discriminator loss={smoothed_value_disc}",
|
515 |
-
"model_training_info_2": "{model_name} | epoch={epoch} | step={global_step} | {epoch_recorder} | lowest value={lowest_value_rounded} (epoch {lowest_value_epoch} and step {lowest_value_step})",
|
516 |
-
"model_training_info_3": "{model_name} | epoch={epoch} | step={global_step} | {epoch_recorder}",
|
517 |
-
"training_error": "An error occurred while training the model:",
|
518 |
-
"separator_info": "Initializing with output path: {output_dir}, output format: {output_format}",
|
519 |
-
"output_dir_is_none": "Output folder not specified. Using current working directory.",
|
520 |
-
">0or=1": "Normalization threshold must be greater than 0 and less than or equal to 1.",
|
521 |
-
"output_single": "Single root output requested; only one file ({output_single_stem}) will be written",
|
522 |
-
"step2": "The second step will be reversed using spectrogram instead of waveform. This may improve quality but is slightly slower.",
|
523 |
-
"other_than_zero": "Sample rate set to {sample_rate} but must be a nonzero integer.",
|
524 |
-
"too_high": "Sample rate set to {sample_rate}. Enter a less ambitious value.",
|
525 |
-
"sr_not_valid": "Sample rate must be a nonzero integer. Please provide a valid integer.",
|
526 |
-
"name_ver": "Version {name}",
|
527 |
-
"os": "Operating System",
|
528 |
-
"platform_info": "System: {system_info} Name: {node} Release: {release} Machine: {machine} Processor: {processor}",
|
529 |
-
"install_ffmpeg": "FFmpeg installed",
|
530 |
-
"none_ffmpeg": "FFmpeg is not installed. Please install FFmpeg to use this package.",
|
531 |
-
"install_onnx": "ONNX Runtime package {pu} installed with version",
|
532 |
-
"running_in_cpu": "Unable to configure hardware acceleration, running in CPU mode",
|
533 |
-
"running_in_cuda": "CUDA available in Torch, setting Torch device to CUDA",
|
534 |
-
"onnx_have": "ONNXruntime available {have}, enabling acceleration",
|
535 |
-
"onnx_not_have": "{have} not available in ONNXruntime; acceleration will NOT be enabled",
|
536 |
-
"python_not_install": "Python package: {package_name} is not installed",
|
537 |
-
"hash": "Calculating hash for model file {model_path}",
|
538 |
-
"ioerror": "IOError while seeking -10 MB or reading model file to compute hash: {e}",
|
539 |
-
"cancel_download": "File already exists at {output_path}, skipping download",
|
540 |
-
"download_model": "Downloading file from {url} to {output_path} with a timeout of 300 seconds",
|
541 |
-
"download_error": "Failed to download file from {url}, response code: {status_code}",
|
542 |
-
"vip_model": "Model: '{model_friendly_name}' is a premium model intended by Anjok07 only for paid subscriber access.",
|
543 |
-
"vip_print": "Hey there, if you haven't subscribed, please consider supporting UVR's developer, Anjok07, by subscribing here: https://patreon.com/uvr",
|
544 |
-
"search_model": "Searching for model {model_filename} in the list of supported models in the group",
|
545 |
-
"load_download_json": "Downloaded model list loaded",
|
546 |
-
"single_model": "Identified single model file: {model_friendly_name}",
|
547 |
-
"not_found_model": "Model not found in the UVR repository, attempting to download from the audio model separation repository...",
|
548 |
-
"single_model_path": "Returning path for single model file: {model_path}",
|
549 |
-
"find_model": "Input file name {model_filename} found in multi-file model: {model_friendly_name}",
|
550 |
-
"find_models": "Identified multi-file model: {model_friendly_name}, iterating through files to download",
|
551 |
-
"find_path": "Attempting to determine download PATH for config pair",
|
552 |
-
"not_found_model_warehouse": "Model not found in the UVR repository, attempting to download from the audio model separation repository...",
|
553 |
-
"yaml_warning": "The model name you specified, {model_filename}, is actually a model config file rather than a model file.",
|
554 |
-
"yaml_warning_2": "We found a model matching this config file: {config_key}, so we’ll use that model file for this run.",
|
555 |
-
"yaml_warning_3": "To avoid confusing/inconsistent behavior in the future, specify the actual model file name instead.",
|
556 |
-
"yaml_debug": "Config YAML model file not found in UVR repository, attempting to download from the audio model separation repository...",
|
557 |
-
"download_model_friendly": "All files downloaded for model {model_friendly_name}, returning original path {model_path}",
|
558 |
-
"not_found_model_2": "Model file {model_filename} not found in the supported files",
|
559 |
-
"load_yaml": "Loading model data from YAML at path {model_data_yaml_filepath}",
|
560 |
-
"load_yaml_2": "Model data loaded from YAML file: {model_data}",
|
561 |
-
"hash_md5": "Computing MD5 hash for model file to identify model parameters from UVR data...",
|
562 |
-
"model_hash": "Model {model_path} has hash {model_hash}",
|
563 |
-
"mdx_data": "MDX model data path set to {mdx_model_data_path}",
|
564 |
-
"load_mdx": "Loading MDX model parameters from UVR model data file...",
|
565 |
-
"model_not_support": "Unsupported model file: no parameters found for MD5 hash {model_hash} in UVR model data for MDX vault.",
|
566 |
-
"uvr_json": "Model data loaded from UVR JSON with hash {model_hash}: {model_data}",
|
567 |
-
"loading_model": "Loading model {model_filename}...",
|
568 |
-
"download_model_friendly_2": "Downloaded model, friendly name: {model_friendly_name}, Model path: {model_path}",
|
569 |
-
"model_type_not_support": "Unsupported model type: {model_type}",
|
570 |
-
"demucs_not_support_python<3.10": "Demucs models require Python version 3.10 or higher.",
|
571 |
-
"import_module": "Importing module for model type",
|
572 |
-
"initialization": "Initializing separator class for model type",
|
573 |
-
"loading_model_success": "Model loading completed.",
|
574 |
-
"loading_model_duration": "Model loading duration",
|
575 |
-
"starting_separator": "Starting separation process for audio file path",
|
576 |
-
"normalization": "Normalization threshold set to {normalization_threshold}, waveform will be scaled down to this maximum amplitude to prevent clipping.",
|
577 |
-
"loading_separator_model": "Downloading model {model_filename}...",
|
578 |
-
"separator_success_3": "Separation process completed.",
|
579 |
-
"separator_duration": "Separation duration",
|
580 |
-
"downloading_model": "Downloaded model, type: {model_type}, friendly name: {model_friendly_name}, Model path: {model_path}, Model data: {model_data_dict_size} items",
|
581 |
-
"not_found_model_signature": "Pre-trained model with signature {sig} not found.",
|
582 |
-
"repo_must_be_folder": "{repo} must exist and be a directory.",
|
583 |
-
"demucs_info": "Demucs parameters: Segment size = {segment_size}, Segment size active = {segments_enabled}",
|
584 |
-
"demucs_info_2": "Demucs parameters: Number of predictions = {shifts}, Overlap = {overlap}",
|
585 |
-
"start_demucs": "Demucs Separator initialization completed",
|
586 |
-
"start_separator": "Starting separation process...",
|
587 |
-
"prepare_mix": "Preparing mixture...",
|
588 |
-
"demix": "Mixture prepared for demixing. Shape: {shape}",
|
589 |
-
"cancel_mix": "Loading model for demixing...",
|
590 |
-
"model_review": "Model loaded and set to evaluation mode.",
|
591 |
-
"del_gpu_cache_after_demix": "Cleared model and GPU cache after demixing.",
|
592 |
-
"process_output_file": "Processing output file...",
|
593 |
-
"source_length": "Processing source array, source length is {source_length}",
|
594 |
-
"process_ver": "Processing source version...",
|
595 |
-
"set_map": "Set source map to {part} parts...",
|
596 |
-
"process_all_part": "Processing for all root parts...",
|
597 |
-
"skip_part": "Skipping root part {stem_name} as out_single_stem is set to {output_single_stem}...",
|
598 |
-
"starting_demix_demucs": "Starting the demix process in demix_demucs...",
|
599 |
-
"model_infer": "Running model inference...",
|
600 |
-
"del_all_but_one": "Duplicate pre-trained models detected for signature {xp_sig}. Please delete all but one.",
|
601 |
-
"name_not_pretrained": "{name} is not a pre-trained model or a model bundle.",
|
602 |
-
"invalid_checksum": "Invalid checksum for file {path}, expected {checksum} but got {actual_checksum}",
|
603 |
-
"mdx_info": "MDX parameters: Batch size = {batch_size}, Segment size = {segment_size}",
|
604 |
-
"mdx_info_2": "MDX parameters: Overlap = {overlap}, Hop_length = {hop_length}, Denoising enabled = {enable_denoise}",
|
605 |
-
"mdx_info_3": "MDX parameters",
|
606 |
-
"load_model_onnx": "Loading ONNX model for inference...",
|
607 |
-
"load_model_onnx_success": "Successfully loaded model using ONNXruntime inference session.",
|
608 |
-
"onnx_to_pytorch": "Model converted from ONNX to PyTorch due to mismatched segment size with dim_t, processing may be slower.",
|
609 |
-
"stft": "Inverse STFT applied. Returning result with shape",
|
610 |
-
"no_denoise": "Model running on spectrum without denoising.",
|
611 |
-
"mix": "Preparing mix for input audio file {audio_file_path}...",
|
612 |
-
"normalization_demix": "Normalizing mix prior to demixing...",
|
613 |
-
"mix_success": "Mix preparation completed.",
|
614 |
-
"primary_source": "Normalizing primary source...",
|
615 |
-
"secondary_source": "Producing secondary source: Mixing in compatible mode",
|
616 |
-
"invert_using_spec": "Inverting secondary stem using spectrum when invert_USE_spec is set to True",
|
617 |
-
"invert_using_spec_2": "Inverting secondary stem by subtracting transformed stem from the initial transformed mix",
|
618 |
-
"enable_denoise": "Model running on both positive and negative spectrums for denoising.",
|
619 |
-
"is_match_mix": "is_match_mix: Predicted spectrum obtained directly from STFT output.",
|
620 |
-
"save_secondary_stem_output_path": "Saving secondary stem {stem_name} to {stem_output_path}...",
|
621 |
-
"starting_model": "Initializing model settings...",
|
622 |
-
"input_info": "Model input parameters",
|
623 |
-
"model_settings": "Model settings",
|
624 |
-
"initialize_mix": "Initializing mix with is_ckpt = {is_ckpt}. Initial mix shape: {shape}",
|
625 |
-
"!=2": "Expected 2-channel audio signal but got {shape} channels",
|
626 |
-
"process_check": "Processing in checkpoint mode...",
|
627 |
-
"stft_2": "STFT applied to mix. Spectrum shape: {shape}",
|
628 |
-
"cache": "Computed padding",
|
629 |
-
"shape": "Mix shape after padding: {shape}, Number of parts: {num_chunks}",
|
630 |
-
"process_no_check": "Processing in no-checkpoint mode...",
|
631 |
-
"n_sample_or_pad": "Number of samples: {n_sample}, Computed padding: {pad}",
|
632 |
-
"shape_2": "Mix shape after padding",
|
633 |
-
"process_part": "Processed part {mix_waves}: Start {i}, End {ii}",
|
634 |
-
"mix_waves_to_tensor": "Converted mix_waves to tensor. Tensor shape: {shape}",
|
635 |
-
"mix_match": "Mix mode Match; applying compensation factor.",
|
636 |
-
"tar_waves": "tar_waves. Shape",
|
637 |
-
"normalization_2": "Normalizing result by dividing it by divisor.",
|
638 |
-
"mix_wave": "Processing mix_wave batch",
|
639 |
-
"mix_or_batch": "Mix parts into batches. Number of batches",
|
640 |
-
"demix_is_match_mix": "Starting demix process with is_match_mix,",
|
641 |
-
"mix_shape": "Root mix parts stored. Shape",
|
642 |
-
"chunk_size_or_overlap": "Chunk size for compatible mixing: {chunk_size}, Overlap: {overlap}",
|
643 |
-
"chunk_size_or_overlap_standard": "Standard chunk size: {chunk_size}, Overlap: {overlap}",
|
644 |
-
"calc_size": "Generated size calculated",
|
645 |
-
"window": "Window applied to this segment.",
|
646 |
-
"process_part_2": "Processing segment {total}/{total_chunks}: Start {start}, End {end}",
|
647 |
-
"all_process_part": "Total segments to process",
|
648 |
-
"step_or_overlap": "Step size to process parts: {step} with overlap set to {overlap}.",
|
649 |
-
"mix_cache": "Mix prepared with padding. Mix shape",
|
650 |
-
"dims": "Cannot use sin/cos position encoding with odd dimensions (dim={dims})",
|
651 |
-
"activation": "activation must be relu/gelu, not {activation}",
|
652 |
-
"length_or_training_length": "Provided length {length} exceeds training duration {training_length}",
|
653 |
-
"type_not_valid": "Invalid type for",
|
654 |
-
"del_parameter": "Removing non-existent parameter ",
|
655 |
-
"info": "Common parameters: Model name = {model_name}, Model path = {model_path}",
|
656 |
-
"info_2": "Common parameters: Output path = {output_dir}, Output format = {output_format}",
|
657 |
-
"info_3": "Common parameters: Normalization threshold = {normalization_threshold}",
|
658 |
-
"info_4": "Common parameters: Denoising enabled = {enable_denoise}, Single stem output = {output_single_stem}",
|
659 |
-
"info_5": "Common parameters: Inversion using specs = {invert_using_spec}, Sample rate = {sample_rate}",
|
660 |
-
"info_6": "Common parameters: Primary root name = {primary_stem_name}, Secondary root name = {secondary_stem_name}",
|
661 |
-
"info_7": "Common parameters: Karaoke mode = {is_karaoke}, BV model = {is_bv_model}, BV model rebalancing = {bv_model_rebalance}",
|
662 |
-
"success_process": "Completed processing root {stem_name} and writing audio...",
|
663 |
-
"load_audio": "Loading audio from file",
|
664 |
-
"load_audio_success": "Audio loaded. Sample rate: {sr}, Audio shape: {shape}",
|
665 |
-
"convert_mix": "Converting provided mix array.",
|
666 |
-
"convert_shape": "Converted mix shape: {shape}",
|
667 |
-
"audio_not_valid": "Audio file {audio_path} is empty or invalid",
|
668 |
-
"audio_valid": "Audio file is valid and contains data.",
|
669 |
-
"mix_single": "Mix is mono. Converting to stereo.",
|
670 |
-
"convert_mix_audio": "Converted to stereo mix.",
|
671 |
-
"mix_success_2": "Mix preparation completed.",
|
672 |
-
"duration": "Audio duration is {duration_hours} hours ({duration_seconds} seconds).",
|
673 |
-
"write": "Using {name} to write.",
|
674 |
-
"write_audio": "Writing {name} with root path:",
|
675 |
-
"original_not_valid": "Warning: Original source array is nearly silent or empty.",
|
676 |
-
"shape_audio": "Audio data shape before processing",
|
677 |
-
"convert_data": "Data type before conversion",
|
678 |
-
"original_source_to_int16": "Converted original_source to int16.",
|
679 |
-
"shape_audio_2": "Interleaved audio data shape",
|
680 |
-
"create_audiosegment": "Successfully created AudioSegment.",
|
681 |
-
"create_audiosegment_error": "Specific error while creating AudioSegment",
|
682 |
-
"export_error": "Error exporting audio file",
|
683 |
-
"export_success": "Successfully exported audio file to",
|
684 |
-
"clean": "Running garbage collection...",
|
685 |
-
"clean_cache": "Clearing {name} cache...",
|
686 |
-
"del_path": "Deleting path, source, and root of input audio file...",
|
687 |
-
"not_success": "Process was not successful: ",
|
688 |
-
"resample_error": "Error during resampling",
|
689 |
-
"shapes": "Shapes",
|
690 |
-
"wav_resolution": "Resolution type",
|
691 |
-
"warnings": "Warning: Extremely aggressive values detected",
|
692 |
-
"warnings_2": "Warning: NaN or infinite values detected in wave input. Shape",
|
693 |
-
"process_file": "Processing file... \n",
|
694 |
-
"save_instruments": "Saving reverse track...",
|
695 |
-
"assert": "Audio files must have the same shape - Mix: {mixshape}, Inst: {instrumentalshape}",
|
696 |
-
"rubberband": "Rubberband CLI cannot be executed. Please ensure Rubberband-CLI is installed.",
|
697 |
-
"rate": "Rate must be strictly positive",
|
698 |
-
"gdown_warning": "You provided a Google Drive link that is not a direct file download link. Consider using the `--fuzzy` option or the following URL: https://drive.google.com/uc?id={file_id}",
|
699 |
-
"gdown_error": "Could not retrieve the public link for the file. You may need to change its permissions to 'Anyone with the link' or there may already be excessive access permissions.",
|
700 |
-
"gdown_error_2": "Failed to retrieve file path:\n\n{error}\n\nYou can still access the file via a browser:\n\n\t{url_origin}\n\nbut Gdown cannot. Please check your connection and permissions.",
|
701 |
-
"temps": "Several temporary files exist for continuation:",
|
702 |
-
"del_all_temps": "Please delete all except one to continue downloading.",
|
703 |
-
"continue": "Continue:",
|
704 |
-
"to": "To:",
|
705 |
-
"gdown_value_error": "A path or ID must be specified",
|
706 |
-
"missing_url": "URL is missing",
|
707 |
-
"mac_not_match": "MAC does not match",
|
708 |
-
"file_not_access": "File is not accessible",
|
709 |
-
"int_resp==-3": "Request failed, retrying"
|
710 |
}
|
|
|
1 |
+
{
|
2 |
+
"set_precision": "Override preprocessing and config.json to use {precision}.",
|
3 |
+
"set_lang": "Display language set to {lang}.",
|
4 |
+
"no_support_gpu": "Unfortunately, no compatible GPU is available to support your training.",
|
5 |
+
"text": "text",
|
6 |
+
"upload_success": "File {name} uploaded successfully.",
|
7 |
+
"download_url": "Download from the link",
|
8 |
+
"download_from_csv": "Download from the CSV model repository",
|
9 |
+
"download_from_applio": "Download model from Applio",
|
10 |
+
"upload": "Upload",
|
11 |
+
"option_not_valid": "Invalid option!",
|
12 |
+
"list_model": "Model list",
|
13 |
+
"success": "Completed!",
|
14 |
+
"index": "index",
|
15 |
+
"model": "model",
|
16 |
+
"zip": "compress",
|
17 |
+
"search": "search",
|
18 |
+
"provide_file": "Please provide a valid {filename} file!",
|
19 |
+
"start": "Starting {start}...",
|
20 |
+
"not_found": "Not found {name}.",
|
21 |
+
"found": "Found {results} results!",
|
22 |
+
"download_music": "download music",
|
23 |
+
"download": "download",
|
24 |
+
"provide_url": "Please provide a link.",
|
25 |
+
"provide_name_is_save": "Please provide a model name to save.",
|
26 |
+
"not_support_url": "Your model link is not supported.",
|
27 |
+
"error_occurred": "An error occurred: {e}.",
|
28 |
+
"not_model": "The file you uploaded is not a model file!",
|
29 |
+
"unable_analyze_model": "Unable to analyze the model!",
|
30 |
+
"credits": "\n**This project is brewed by [Phạm Huỳnh Anh]({author})**\n\n**The project is built based on several major projects such as:**\n\n**Conversion, Processing, Extraction, Training, Model Reading, Model Fusion, Training Model, Model Repository...: [Applio]({applio}) by the [AI Hispano]({ai_hispano}) group**\n\n**Extraction Methods, Information Display, Logging, Training Models...: [Retrieval-based-Voice-Conversion-WebUI]({rvc_webui}) by author [RVC BOSS]({rvc_boss})**\n\n**Music Separation Models MDX-Net and Demucs: [Python-audio-separator]({python_audio_separator}) by author [Andrew Beveridge]({andrew_beveridge})**\n",
|
31 |
+
"download_pretrain": "Downloading pre-trained model...",
|
32 |
+
"provide_pretrain": "Please provide a pre-trained model link {dg}.",
|
33 |
+
"provide_hubert": "Please provide a link to the embedding model.",
|
34 |
+
"sr_not_same": "The sample rates of the two models are not the same.",
|
35 |
+
"architectures_not_same": "Cannot merge models. The architectures are not the same.",
|
36 |
+
"fushion_model": "model fusion",
|
37 |
+
"model_fushion_info": "The model {name} is fused from {pth_1} and {pth_2} with a ratio of {ratio}.",
|
38 |
+
"not_found_create_time": "Creation time not found.",
|
39 |
+
"format_not_valid": "Invalid format.",
|
40 |
+
"read_info": "Models trained on different applications may produce different information or may not be readable!",
|
41 |
+
"epoch": "epoch.",
|
42 |
+
"step": "step",
|
43 |
+
"sr": "sample rate.",
|
44 |
+
"f0": "pitch training",
|
45 |
+
"version": "version.",
|
46 |
+
"not_f0": "Pitch training not performed",
|
47 |
+
"trained_f0": "Pitch training performed",
|
48 |
+
"model_info": "Model Name: {model_name}\n\n Model Creator: {model_author}\n\nEpoch: {epochs}\n\nSteps: {steps}\n\nVersion: {version}\n\nSample Rate: {sr}\n\nPitch Training: {pitch_guidance}\n\nHash (ID): {model_hash}\n\nCreation Time: {creation_date_str}\n",
|
49 |
+
"input_not_valid": "Please provide valid input!",
|
50 |
+
"output_not_valid": "Please provide valid output!",
|
51 |
+
"apply_effect": "apply effect",
|
52 |
+
"enter_the_text": "Please enter the text to read!",
|
53 |
+
"choose_voice": "Please choose a voice!",
|
54 |
+
"convert": "Converting {name}...",
|
55 |
+
"separator_music": "music separation",
|
56 |
+
"notfound": "Not found",
|
57 |
+
"turn_on_use_audio": "Please enable using separated audio to proceed",
|
58 |
+
"turn_off_convert_backup": "Disable backup voice conversion to use the original voice",
|
59 |
+
"turn_off_merge_backup": "Disable merging backup voice to use the original voice",
|
60 |
+
"not_found_original_vocal": "Original vocal not found!",
|
61 |
+
"convert_vocal": "Converting voice...",
|
62 |
+
"convert_success": "Voice conversion completed!",
|
63 |
+
"convert_backup": "Converting backup voice...",
|
64 |
+
"convert_backup_success": "Backup voice conversion completed!",
|
65 |
+
"merge_backup": "Merging main voice with backup voice...",
|
66 |
+
"merge_success": "Merge completed.",
|
67 |
+
"is_folder": "Input is a folder: Converting all audio files in the folder...",
|
68 |
+
"not_found_in_folder": "No audio files found in the folder!",
|
69 |
+
"batch_convert": "Batch conversion in progress...",
|
70 |
+
"batch_convert_success": "Batch conversion successful!",
|
71 |
+
"create": "create",
|
72 |
+
"provide_name": "Please provide a model name.",
|
73 |
+
"not_found_data": "Data not found",
|
74 |
+
"not_found_data_preprocess": "Processed audio data not found, please reprocess.",
|
75 |
+
"not_found_data_extract": "Extracted audio data not found, please re-extract.",
|
76 |
+
"provide_pretrained": "Please provide pre-trained {dg}.",
|
77 |
+
"download_pretrained": "Download pre-trained {dg}{rvc_version} original",
|
78 |
+
"not_found_pretrain": "Pre-trained {dg} not found",
|
79 |
+
"not_use_pretrain": "No pre-trained model will be used",
|
80 |
+
"training": "training",
|
81 |
+
"display_title": "<h1> 🎵 Voice model conversion and training interface created by Anh 🎵 <h1>",
|
82 |
+
"rick_roll": "Click here if you want to be Rick Roll :) ---> [RickRoll]({rickroll})",
|
83 |
+
"terms_of_use": "**Please do not use the project for any unethical, illegal, or harmful purposes to individuals or organizations...**",
|
84 |
+
"exemption": "**In cases where users do not comply with the terms or violate them, I will not be responsible for any claims, damages, or liabilities, whether in contract, negligence, or other causes arising from, outside of, or related to the software, its use, or other transactions associated with it.**",
|
85 |
+
"separator_tab": "Music Separation",
|
86 |
+
"4_part": "A simple music separation system can separate into 4 parts: Music, vocals, main vocals, backup vocals",
|
87 |
+
"clear_audio": "Clean audio",
|
88 |
+
"separator_backing": "Separate backup vocals",
|
89 |
+
"denoise_backing": "Denoise backup vocals",
|
90 |
+
"denoise_mdx": "Denoise MDX separation",
|
91 |
+
"use_mdx": "Use MDX",
|
92 |
+
"dereveb_audio": "Remove vocal reverb",
|
93 |
+
"dereveb_backing": "Remove backup reverb",
|
94 |
+
"denoise_dereveb": "Denoise reverb",
|
95 |
+
"separator_model": "Music separation model",
|
96 |
+
"separator_backing_model": "Backup separation model",
|
97 |
+
"shift": "Number of predictions",
|
98 |
+
"shift_info": "Higher is better quality but slower and uses more resources",
|
99 |
+
"segments_size": "Segment Size",
|
100 |
+
"segments_size_info": "Higher is better quality but uses more resources",
|
101 |
+
"batch_size": "Batch size",
|
102 |
+
"batch_size_info": "Number of samples processed simultaneously in one training cycle. Higher can cause memory overflow",
|
103 |
+
"mdx_batch_size_info": "Number of samples processed at a time. Batch processing optimizes calculations. Large batches can cause memory overflow; small batches reduce resource efficiency",
|
104 |
+
"overlap": "Overlap",
|
105 |
+
"overlap_info": "Overlap amount between prediction windows",
|
106 |
+
"export_format": "Audio format",
|
107 |
+
"export_info": "The audio format to export the audio file in",
|
108 |
+
"output_separator": "Separated audio",
|
109 |
+
"hop_length_info": "Indicates the time interval for shifting the analysis window on the audio signal when performing transformations. Smaller values increase detail but require more computational resources",
|
110 |
+
"drop_audio": "Drop audio here",
|
111 |
+
"drop_text": "Drop text file here",
|
112 |
+
"use_url": "Use YouTube link",
|
113 |
+
"url_audio": "Link to audio",
|
114 |
+
"downloads": "Downloads",
|
115 |
+
"clean_strength": "Audio cleaning strength",
|
116 |
+
"clean_strength_info": "Strength of the audio cleaner for filtering vocals during export",
|
117 |
+
"input_output": "Audio input, output",
|
118 |
+
"audio_path": "Input audio path",
|
119 |
+
"refesh": "Refresh",
|
120 |
+
"output_folder": "Output audio folder path",
|
121 |
+
"output_folder_info": "Enter the folder path where the audio will be exported",
|
122 |
+
"input_audio": "Audio input",
|
123 |
+
"instruments": "Instruments",
|
124 |
+
"original_vocal": "Original vocal",
|
125 |
+
"main_vocal": "Main vocal",
|
126 |
+
"backing_vocal": "Backup vocal",
|
127 |
+
"convert_audio": "Convert Audio",
|
128 |
+
"convert_info": "Convert audio using a trained voice model",
|
129 |
+
"upscale_audio": "Upscale quality",
|
130 |
+
"autotune": "Auto-tune",
|
131 |
+
"use_audio": "Use separated audio",
|
132 |
+
"convert_original": "Convert original voice",
|
133 |
+
"convert_backing": "Convert backup voice",
|
134 |
+
"not_merge_backing": "Do not merge backup voice",
|
135 |
+
"merge_instruments": "Merge instruments",
|
136 |
+
"pitch": "Pitch",
|
137 |
+
"pitch_info": "Recommendation: set to 12 to change male voice to female and vice versa",
|
138 |
+
"model_accordion": "Model and index",
|
139 |
+
"model_name": "Model file",
|
140 |
+
"index_path": "Index file",
|
141 |
+
"index_strength": "Index strength",
|
142 |
+
"index_strength_info": "Higher values increase strength. However, lower values may reduce artificial effects in the audio",
|
143 |
+
"output_path": "Audio output path",
|
144 |
+
"output_path_info": "Enter the output path (leave it as .wav format; it will auto-correct during conversion)",
|
145 |
+
"setting": "General settings",
|
146 |
+
"f0_method": "Extraction method",
|
147 |
+
"f0_method_info": "Method used for data extraction",
|
148 |
+
"f0_method_hybrid": "HYBRID extraction method",
|
149 |
+
"f0_method_hybrid_info": "Combines two different extraction types",
|
150 |
+
"hubert_model": "Embedding model",
|
151 |
+
"hubert_info": "Pre-trained model to assist embedding",
|
152 |
+
"modelname": "Model name",
|
153 |
+
"modelname_info": "If you have your own model, just upload it and input the name here",
|
154 |
+
"split_audio": "Split audio",
|
155 |
+
"split_audio_info": "Splitting audio into smaller parts for conversion can speed up the process",
|
156 |
+
"batch_process": "Batch processing",
|
157 |
+
"batch_process_info": "Batch processing can reduce training time but may overload resources",
|
158 |
+
"autotune_rate": "Auto-tune level",
|
159 |
+
"autotune_rate_info": "Level of auto-tuning adjustment",
|
160 |
+
"resample": "Resample",
|
161 |
+
"resample_info": "Resample post-processing to the final sample rate; 0 means no resampling",
|
162 |
+
"filter_radius": "Median filter",
|
163 |
+
"filter_radius_info": "If greater than three, median filtering is applied. The value represents the filter radius and can reduce breathiness or noise.",
|
164 |
+
"volume_envelope": "Volume envelope",
|
165 |
+
"volume_envelope_info": "Use the input volume envelope to replace or mix with the output volume envelope. The closer to 1, the more the output envelope is used",
|
166 |
+
"protect": "Consonant protection",
|
167 |
+
"protect_info": "Protect distinct consonants and breathing sounds to prevent audio tearing and other artifacts. Increasing this value provides comprehensive protection. Reducing it may reduce protection but also minimize indexing effects",
|
168 |
+
"output_convert": "Converted audio",
|
169 |
+
"main_convert": "Convert main voice",
|
170 |
+
"main_or_backing": "Main voice + Backup voice",
|
171 |
+
"voice_or_instruments": "Voice + Instruments",
|
172 |
+
"convert_text": "Convert Text",
|
173 |
+
"convert_text_markdown": "## Convert Text to Speech",
|
174 |
+
"convert_text_markdown_2": "Convert text to speech and read aloud using the trained voice model",
|
175 |
+
"input_txt": "Input data from a text file (.txt)",
|
176 |
+
"text_to_speech": "Text to read",
|
177 |
+
"voice_speed": "Reading speed",
|
178 |
+
"voice_speed_info": "Speed of the voice",
|
179 |
+
"tts_1": "1. Convert Text",
|
180 |
+
"tts_2": "2. Convert Speech",
|
181 |
+
"voice": "Voices by country",
|
182 |
+
"output_tts": "Output speech path",
|
183 |
+
"output_tts_convert": "Converted speech output path",
|
184 |
+
"tts_output": "Enter the output path",
|
185 |
+
"output_tts_markdown": "Unconverted and converted audio",
|
186 |
+
"output_text_to_speech": "Generated speech from text-to-speech conversion",
|
187 |
+
"output_file_tts_convert": "Speech converted using the model",
|
188 |
+
"output_audio": "Audio output",
|
189 |
+
"provide_output": "Enter the output path",
|
190 |
+
"audio_effects": "Audio Effects",
|
191 |
+
"apply_audio_effects": "## Add Additional Audio Effects",
|
192 |
+
"audio_effects_edit": "Edit and add effects to audio",
|
193 |
+
"reverb": "Reverb effect",
|
194 |
+
"chorus": "Chorus effect",
|
195 |
+
"delay": "Delay effect",
|
196 |
+
"more_option": "Additional options",
|
197 |
+
"phaser": "Phaser effect",
|
198 |
+
"compressor": "Compressor effect",
|
199 |
+
"apply": "Apply",
|
200 |
+
"reverb_freeze": "Freeze mode",
|
201 |
+
"reverb_freeze_info": "Create a continuous echo effect when this mode is enabled",
|
202 |
+
"room_size": "Room size",
|
203 |
+
"room_size_info": "Adjust the room space to create reverberation",
|
204 |
+
"damping": "Damping",
|
205 |
+
"damping_info": "Adjust the level of absorption to control the amount of reverberation",
|
206 |
+
"wet_level": "Reverb signal level",
|
207 |
+
"wet_level_info": "Adjust the level of the reverb signal effect",
|
208 |
+
"dry_level": "Original signal level",
|
209 |
+
"dry_level_info": "Adjust the level of the signal without effects",
|
210 |
+
"width": "Audio width",
|
211 |
+
"width_info": "Adjust the width of the audio space",
|
212 |
+
"chorus_depth": "Chorus depth",
|
213 |
+
"chorus_depth_info": "Adjust the intensity of the chorus to create a wider sound",
|
214 |
+
"chorus_rate_hz": "Frequency",
|
215 |
+
"chorus_rate_hz_info": "Adjust the oscillation speed of the chorus effect",
|
216 |
+
"chorus_mix": "Mix signals",
|
217 |
+
"chorus_mix_info": "Adjust the mix level between the original and the processed signal",
|
218 |
+
"chorus_centre_delay_ms": "Center delay (ms)",
|
219 |
+
"chorus_centre_delay_ms_info": "The delay time between stereo channels to create the chorus effect",
|
220 |
+
"chorus_feedback": "Feedback",
|
221 |
+
"chorus_feedback_info": "Adjust the amount of the effect signal fed back into the original signal",
|
222 |
+
"delay_seconds": "Delay time",
|
223 |
+
"delay_seconds_info": "Adjust the delay time between the original and the processed signal",
|
224 |
+
"delay_feedback": "Delay feedback",
|
225 |
+
"delay_feedback_info": "Adjust the amount of feedback signal, creating a repeating effect",
|
226 |
+
"delay_mix": "Delay signal mix",
|
227 |
+
"delay_mix_info": "Adjust the mix level between the original and delayed signal",
|
228 |
+
"fade": "Fade effect",
|
229 |
+
"bass_or_treble": "Bass and treble",
|
230 |
+
"limiter": "Threshold limiter",
|
231 |
+
"distortion": "Distortion effect",
|
232 |
+
"gain": "Audio gain",
|
233 |
+
"bitcrush": "Bit reduction effect",
|
234 |
+
"clipping": "Clipping effect",
|
235 |
+
"fade_in": "Fade-in effect (ms)",
|
236 |
+
"fade_in_info": "Time for the audio to gradually increase from 0 to normal level",
|
237 |
+
"fade_out": "Fade-out effect (ms)",
|
238 |
+
"fade_out_info": "Time for the audio to gradually decrease to 0",
|
239 |
+
"bass_boost": "Bass boost level (dB)",
|
240 |
+
"bass_boost_info": "Boost level for bass frequencies",
|
241 |
+
"bass_frequency": "Low-pass filter cutoff frequency (Hz)",
|
242 |
+
"bass_frequency_info": "Frequency below which the sound will start to reduce. Lower frequencies make bass clearer",
|
243 |
+
"treble_boost": "Treble boost level (dB)",
|
244 |
+
"treble_boost_info": "Boost level for high frequencies",
|
245 |
+
"treble_frequency": "High-pass filter cutoff frequency (Hz)",
|
246 |
+
"treble_frequency_info": "Sounds below this frequency will be filtered out. Higher frequencies retain more treble",
|
247 |
+
"limiter_threashold_db": "Limiter threshold",
|
248 |
+
"limiter_threashold_db_info": "Limit the maximum audio level to prevent it from exceeding the threshold",
|
249 |
+
"limiter_release_ms": "Release time",
|
250 |
+
"limiter_release_ms_info": "Time for the audio to return after being limited",
|
251 |
+
"distortion_info": "Adjust the level of distortion to create a noisy effect",
|
252 |
+
"gain_info": "Adjust the volume level of the signal",
|
253 |
+
"clipping_threashold_db": "Clipping threshold",
|
254 |
+
"clipping_threashold_db_info": "Trim signals exceeding the threshold, creating a distorted sound",
|
255 |
+
"bitcrush_bit_depth": "Bit depth",
|
256 |
+
"bitcrush_bit_depth_info": "Reduce audio quality by decreasing bit depth, creating a distorted effect",
|
257 |
+
"phaser_depth": "Phaser depth",
|
258 |
+
"phaser_depth_info": "Adjust the depth of the effect, impacting its intensity",
|
259 |
+
"phaser_rate_hz": "Frequency",
|
260 |
+
"phaser_rate_hz_info": "Adjust the frequency of the phaser effect",
|
261 |
+
"phaser_mix": "Mix signal",
|
262 |
+
"phaser_mix_info": "Adjust the mix level between the original and processed signals",
|
263 |
+
"phaser_centre_frequency_hz": "Center frequency",
|
264 |
+
"phaser_centre_frequency_hz_info": "The center frequency of the phaser effect, affecting the adjusted frequencies",
|
265 |
+
"phaser_feedback": "Feedback",
|
266 |
+
"phaser_feedback_info": "Adjust the feedback level of the effect, creating a stronger or lighter phaser feel",
|
267 |
+
"compressor_threashold_db": "Compressor threshold",
|
268 |
+
"compressor_threashold_db_info": "The threshold level above which the audio will be compressed",
|
269 |
+
"compressor_ratio": "Compression ratio",
|
270 |
+
"compressor_ratio_info": "Adjust the level of audio compression when exceeding the threshold",
|
271 |
+
"compressor_attack_ms": "Attack time (ms)",
|
272 |
+
"compressor_attack_ms_info": "Time for compression to start taking effect after the audio exceeds the threshold",
|
273 |
+
"compressor_release_ms": "Release time",
|
274 |
+
"compressor_release_ms_info": "Time for the audio to return to normal after being compressed",
|
275 |
+
"create_dataset_url": "Link to audio (use commas for multiple links)",
|
276 |
+
"createdataset": "Create dataset",
|
277 |
+
"create_dataset_markdown": "## Create Training Data from YouTube",
|
278 |
+
"create_dataset_markdown_2": "Process and create training datasets using YouTube links",
|
279 |
+
"denoise": "Model denoise",
|
280 |
+
"skip": "Skip seconds",
|
281 |
+
"model_ver": "Voice separation version",
|
282 |
+
"model_ver_info": "The model version for separating vocals",
|
283 |
+
"create_dataset_info": "Dataset creation information",
|
284 |
+
"output_data": "Dataset output",
|
285 |
+
"output_data_info": "Output data after creation",
|
286 |
+
"skip_start": "Skip beginning",
|
287 |
+
"skip_start_info": "Skip the initial seconds of the audio; use commas for multiple audios",
|
288 |
+
"skip_end": "Skip end",
|
289 |
+
"skip_end_info": "Skip the final seconds of the audio; use commas for multiple audios",
|
290 |
+
"training_model": "Train Model",
|
291 |
+
"training_markdown": "Train and build a voice model with a set of voice data",
|
292 |
+
"training_model_name": "Name of the model during training (avoid special characters or spaces)",
|
293 |
+
"sample_rate": "Sample rate",
|
294 |
+
"sample_rate_info": "Sample rate of the model",
|
295 |
+
"training_version": "Model version",
|
296 |
+
"training_version_info": "Version of the model during training",
|
297 |
+
"training_pitch": "Pitch training",
|
298 |
+
"training_pitch_info": "Train the pitch for the model",
|
299 |
+
"upload_dataset": "Upload training data",
|
300 |
+
"preprocess_split": "Should be disabled if data has already been processed",
|
301 |
+
"preprocess_effect": "Processing effects",
|
302 |
+
"preprocess_effect_info": "Should be disabled if data has already been processed",
|
303 |
+
"clear_dataset": "Clean dataset",
|
304 |
+
"clear_dataset_info": "Clean data segments",
|
305 |
+
"preprocess_info": "Preprocessing information",
|
306 |
+
"preprocess_button": "1. Process data",
|
307 |
+
"extract_button": "2. Extract data",
|
308 |
+
"extract_info": "Data extraction information",
|
309 |
+
"total_epoch": "Total epochs",
|
310 |
+
"total_epoch_info": "Total training epochs",
|
311 |
+
"save_epoch": "Save frequency",
|
312 |
+
"save_epoch_info": "Frequency of saving the model during training to allow retraining",
|
313 |
+
"create_index": "Create index",
|
314 |
+
"index_algorithm": "Index algorithm",
|
315 |
+
"index_algorithm_info": "Algorithm for creating the index",
|
316 |
+
"custom_dataset": "Custom folder",
|
317 |
+
"custom_dataset_info": "Custom folder for training data",
|
318 |
+
"overtraining_detector": "Overtraining detector",
|
319 |
+
"overtraining_detector_info": "Check for overtraining during model training",
|
320 |
+
"sync_graph": "Synchronize graph",
|
321 |
+
"sync_graph_info": "Only enable if you need to retrain the model from scratch.",
|
322 |
+
"cache_in_gpu": "Cache model in GPU",
|
323 |
+
"cache_in_gpu_info": "Store the model in GPU cache memory",
|
324 |
+
"dataset_folder": "Folder containing data",
|
325 |
+
"threshold": "Overtraining threshold",
|
326 |
+
"setting_cpu_gpu": "CPU/GPU settings",
|
327 |
+
"gpu_number": "Number of GPUs used",
|
328 |
+
"gpu_number_info": "Number of GPUs used during training",
|
329 |
+
"save_only_latest": "Save only the latest",
|
330 |
+
"save_only_latest_info": "Save only the latest D and G models",
|
331 |
+
"save_every_weights": "Save all models",
|
332 |
+
"save_every_weights_info": "Save all models after each epoch",
|
333 |
+
"gpu_info": "GPU information",
|
334 |
+
"gpu_info_2": "Information about the GPU used during training",
|
335 |
+
"cpu_core": "Number of CPU cores available",
|
336 |
+
"cpu_core_info": "Number of CPU cores used during training",
|
337 |
+
"not_use_pretrain_2": "Do not use pretraining",
|
338 |
+
"not_use_pretrain_info": "Do not use pre-trained models",
|
339 |
+
"custom_pretrain": "Custom pretraining",
|
340 |
+
"custom_pretrain_info": "Customize pre-training settings",
|
341 |
+
"pretrain_file": "Pre-trained model file {dg}",
|
342 |
+
"train_info": "Training information",
|
343 |
+
"export_model": "5. Export Model",
|
344 |
+
"zip_model": "2. Compress model",
|
345 |
+
"output_zip": "Output file after compression",
|
346 |
+
"refesh_pretrain": "Refresh pre-trained model",
|
347 |
+
"model_path": "Model path",
|
348 |
+
"model_ratio": "Model ratio",
|
349 |
+
"model_ratio_info": "Adjusting towards one side will make the model more like that side",
|
350 |
+
"output_model_path": "Model output path",
|
351 |
+
"fushion": "Model Fusion",
|
352 |
+
"fushion_markdown": "## Merge Two Models",
|
353 |
+
"fushion_markdown_2": "Combine two voice models into a single model",
|
354 |
+
"read_model": "Read Information",
|
355 |
+
"read_model_markdown": "## Read Model Information",
|
356 |
+
"read_model_markdown_2": "Retrieve recorded information within the model",
|
357 |
+
"drop_model": "Drop model here",
|
358 |
+
"readmodel": "Read model",
|
359 |
+
"model_path_info": "Enter the path to the model file",
|
360 |
+
"modelinfo": "Model Information",
|
361 |
+
"download_markdown": "## Download Model",
|
362 |
+
"download_markdown_2": "Download voice models, pre-trained models, and embedding models",
|
363 |
+
"model_download": "Download voice model",
|
364 |
+
"model_url": "Link to the model",
|
365 |
+
"30s": "Please wait about 30 seconds. The system will restart automatically!",
|
366 |
+
"fp_select": "Setting precision...",
|
367 |
+
"fp_select_2": "Precision set to: {fp}",
|
368 |
+
"model_download_select": "Choose a model download method",
|
369 |
+
"model_warehouse": "Model repository",
|
370 |
+
"get_model": "Retrieve model",
|
371 |
+
"name_to_search": "Name to search",
|
372 |
+
"search_2": "Search",
|
373 |
+
"select_download_model": "Choose a searched model (Click to select)",
|
374 |
+
"download_pretrainec": "Download pre-trained model",
|
375 |
+
"only_huggingface": "Supports only huggingface.co",
|
376 |
+
"pretrained_url": "Pre-trained model link {dg}",
|
377 |
+
"select_pretrain": "Choose pre-trained model",
|
378 |
+
"select_pretrain_info": "Choose a pre-trained model to download",
|
379 |
+
"pretrain_sr": "Model sample rate",
|
380 |
+
"drop_pretrain": "Drop pre-trained model {dg} here",
|
381 |
+
"hubert_download": "Download embedding model",
|
382 |
+
"hubert_url": "Link to embedding model",
|
383 |
+
"drop_hubert": "Drop embedding model here",
|
384 |
+
"source": "Origin",
|
385 |
+
"source_info": "## Project Origin and Author",
|
386 |
+
"settings": "Settings",
|
387 |
+
"settings_markdown": "## Additional Settings",
|
388 |
+
"settings_markdown_2": "Customize additional features of the project",
|
389 |
+
"lang": "Language",
|
390 |
+
"lang_restart": "The display language in the project (When changing the language, the system will automatically restart after 30 seconds to update)",
|
391 |
+
"change_lang": "Change Language",
|
392 |
+
"fp_train": "Training precision",
|
393 |
+
"fp_info": "Use fp32 for higher precision but requires more memory and strong computation. Use fp16 to reduce memory usage and increase processing speed.",
|
394 |
+
"fp_button": "Change Precision",
|
395 |
+
"theme": "Theme",
|
396 |
+
"theme_restart": "Theme type displayed in the interface (When changing the theme, the system will automatically restart after 30 seconds to update)",
|
397 |
+
"theme_button": "Change Theme",
|
398 |
+
"change_light_dark": "Switch Light/Dark Mode",
|
399 |
+
"tensorboard_url": "Graph URL",
|
400 |
+
"errors_loading_audio": "Error loading audio: {e}",
|
401 |
+
"apply_error": "An error occurred while applying effects: {e}",
|
402 |
+
"indexpath": "Index path",
|
403 |
+
"skip_file": "Part {i} skipped because it is too short: {chunk}ms",
|
404 |
+
"split_total": "Total parts split",
|
405 |
+
"process_audio_error": "An error occurred while processing the audio",
|
406 |
+
"merge_error": "An error occurred while merging audio",
|
407 |
+
"not_found_convert_file": "Processed file not found",
|
408 |
+
"convert_batch": "Batch conversion...",
|
409 |
+
"found_audio": "Found {audio_files} audio files for conversion.",
|
410 |
+
"not_found_audio": "No audio files found!",
|
411 |
+
"error_convert": "An error occurred during audio conversion: {e}",
|
412 |
+
"error_convert_batch": "An error occurred during the conversion of audio segments: {e}",
|
413 |
+
"error_convert_batch_2": "An error occurred during batch audio conversion: {e}",
|
414 |
+
"convert_batch_success": "Batch conversion completed successfully in {elapsed_time} seconds. {output_path}",
|
415 |
+
"convert_audio_success": "File {input_path} converted successfully in {elapsed_time} seconds. {output_path}",
|
416 |
+
"hybrid_methods": "Estimating f0 pitch using methods {methods}",
|
417 |
+
"method_not_valid": "Invalid method",
|
418 |
+
"read_faiss_index_error": "An error occurred while reading the FAISS index: {e}",
|
419 |
+
"read_model_error": "Failed to load model: {e}",
|
420 |
+
"denoise_error": "An error occurred while denoising: {e}",
|
421 |
+
"change_format": "Converting audio to format {output_format}...",
|
422 |
+
"change_format_error": "An error occurred while converting audio format: {e}",
|
423 |
+
"starting_download": "Starting download",
|
424 |
+
"version_not_valid": "Invalid vocal separation version",
|
425 |
+
"create_dataset_value_not_valid": "Enable music separation option to use reverb removal options",
|
426 |
+
"skip<audio": "Cannot skip as skip time is less than audio file length",
|
427 |
+
"skip>audio": "Cannot skip as skip time is greater than audio file length",
|
428 |
+
"=<0": "Skip time is less than or equal to 0 and has been skipped",
|
429 |
+
"skip_warning": "Skip duration ({seconds} seconds) exceeds audio length ({total_duration} seconds). Skipping.",
|
430 |
+
"download_success": "Download completed successfully",
|
431 |
+
"create_dataset_error": "An error occurred while creating the training dataset",
|
432 |
+
"create_dataset_success": "Training dataset creation completed in {elapsed_time} seconds",
|
433 |
+
"skip_start_audio": "Successfully skipped start of audio: {input_file}",
|
434 |
+
"skip_end_audio": "Successfully skipped end of audio: {input_file}",
|
435 |
+
"merge_audio": "Merged all parts containing audio",
|
436 |
+
"separator_process": "Separating vocals: {input}...",
|
437 |
+
"not_found_main_vocal": "Main vocal not found!",
|
438 |
+
"not_found_backing_vocal": "Backup vocal not found!",
|
439 |
+
"not_found_instruments": "Instruments not found",
|
440 |
+
"merge_instruments_process": "Merging vocals with instruments...",
|
441 |
+
"dereverb": "Removing vocal reverb",
|
442 |
+
"dereverb_success": "Successfully removed vocal reverb",
|
443 |
+
"save_index": "Index file saved",
|
444 |
+
"create_index_error": "An error occurred while creating the index",
|
445 |
+
"sr_not_16000": "Sample rate must be 16000",
|
446 |
+
"gpu_not_valid": "Invalid GPU index. Switching to CPU.",
|
447 |
+
"extract_file_error": "An error occurred while extracting the file",
|
448 |
+
"extract_f0_method": "Starting pitch extraction using {num_processes} cores with method {f0_method}...",
|
449 |
+
"extract_f0": "Pitch Extraction",
|
450 |
+
"extract_f0_success": "Pitch extraction completed in {elapsed_time} seconds.",
|
451 |
+
"NaN": "contains NaN values and will be ignored.",
|
452 |
+
"start_extract_hubert": "Starting Hubert embedding extraction...",
|
453 |
+
"not_found_audio_file": "Audio file not found. Please ensure you provided the correct audio.",
|
454 |
+
"extract_hubert": "Extract Hubert embeddings",
|
455 |
+
"process_error": "An error occurred during processing",
|
456 |
+
"extract_hubert_success": "Hubert embedding extraction completed in {elapsed_time} seconds.",
|
457 |
+
"export_process": "Model path",
|
458 |
+
"extract_error": "An error occurred during data extraction",
|
459 |
+
"extract_success": "Data extraction successful",
|
460 |
+
"min_length>=min_interval>=hop_size": "min_length must be greater than or equal to min_interval and hop_size",
|
461 |
+
"max_sil_kept>=hop_size": "max_sil_kept must be greater than or equal to hop_size",
|
462 |
+
"start_preprocess": "Starting data preprocessing with {num_processes} cores...",
|
463 |
+
"not_integer": "Voice ID folder must be an integer; instead got",
|
464 |
+
"preprocess": "Preprocessing Data",
|
465 |
+
"preprocess_success": "Preprocessing completed in {elapsed_time} seconds.",
|
466 |
+
"preprocess_model_success": "Preprocessing data for the model completed successfully",
|
467 |
+
"turn_on_dereverb": "Reverb removal for backup vocals requires enabling reverb removal",
|
468 |
+
"turn_on_separator_backing": "Backup vocal separation requires enabling vocal separation",
|
469 |
+
"demucs_model": "Demucs model",
|
470 |
+
"backing_model_ver": "Backup vocal separation model version",
|
471 |
+
"mdx_model": "MDX model",
|
472 |
+
"clean_audio_success": "Audio cleaned successfully!",
|
473 |
+
"separator_error": "An error occurred during music separation",
|
474 |
+
"separator_success": "Music separation completed in {elapsed_time} seconds",
|
475 |
+
"separator_process_2": "Processing music separation",
|
476 |
+
"separator_success_2": "Music separation successful!",
|
477 |
+
"separator_process_backing": "Processing backup vocal separation",
|
478 |
+
"separator_process_backing_success": "Backup vocal separation successful!",
|
479 |
+
"process_original": "Processing original vocal reverb removal...",
|
480 |
+
"process_original_success": "Original vocal reverb removal successful!",
|
481 |
+
"process_main": "Processing main vocal reverb removal...",
|
482 |
+
"process_main_success": "Main vocal reverb removal successful!",
|
483 |
+
"process_backing": "Processing backup vocal reverb removal...",
|
484 |
+
"process_backing_success": "Backup vocal reverb removal successful!",
|
485 |
+
"save_every_epoch": "Save model after: {save_every_epoch} epochs",
|
486 |
+
"total_e": "Total training epochs: {total_epoch} epochs",
|
487 |
+
"dorg": "Pre-trained G: {pretrainG} | Pre-trained D: {pretrainD}",
|
488 |
+
"training_f0": "Pitch training",
|
489 |
+
"not_gpu": "No GPU detected, reverting to CPU (not recommended)",
|
490 |
+
"sync": "Graph synchronization is now enabled! When synchronization is active, the model will undergo a single training phase. After graphs are synchronized, training will resume with previously specified epochs.",
|
491 |
+
"sync_success": "Graph synchronization successful!",
|
492 |
+
"not_found_checkpoint": "Checkpoint file not found: {checkpoint_path}",
|
493 |
+
"save_checkpoint": "Reloaded checkpoint '{checkpoint_path}' (epoch {checkpoint_dict})",
|
494 |
+
"save_model": "Saved model '{checkpoint_path}' (epoch {iteration})",
|
495 |
+
"sid_error": "Error converting speaker ID '{sid}' to integer. Exception: {e}",
|
496 |
+
"sr_does_not_match": "{sample_rate} Sample rate does not match target {sample_rate2} Sample rate",
|
497 |
+
"spec_error": "An error occurred while retrieving specifications from {spec_filename}: {e}",
|
498 |
+
"time_or_speed_training": "time={current_time} | training speed={elapsed_time_str}",
|
499 |
+
"savemodel": "Saved model '{model_dir}' (epoch {epoch} and step {step})",
|
500 |
+
"model_author": "Credit model to {model_author}",
|
501 |
+
"unregistered": "Model unregistered",
|
502 |
+
"not_author": "Model not credited",
|
503 |
+
"training_author": "Model creator name",
|
504 |
+
"training_author_info": "To credit the model, enter your name here",
|
505 |
+
"extract_model_error": "An error occurred while extracting the model",
|
506 |
+
"start_training": "Starting training",
|
507 |
+
"import_pretrain": "Loaded pre-trained model ({dg}) '{pretrain}'",
|
508 |
+
"not_using_pretrain": "No pre-trained model ({dg}) will be used",
|
509 |
+
"training_warning": "WARNING: Generated loss is lower than the lower threshold loss for the next epoch.",
|
510 |
+
"overtraining_find": "Overtraining detected at epoch {epoch} with smoothed generator loss {smoothed_value_gen} and smoothed discriminator loss {smoothed_value_disc}",
|
511 |
+
"best_epoch": "New best epoch {epoch} with smoothed generator loss {smoothed_value_gen} and smoothed discriminator loss {smoothed_value_disc}",
|
512 |
+
"success_training": "Training completed with {epoch} epochs, {global_step} steps, and {loss_gen_all} total generator loss.",
|
513 |
+
"training_info": "Lowest generator loss: {lowest_value_rounded} at epoch {lowest_value_epoch}, step {lowest_value_step}",
|
514 |
+
"model_training_info": "{model_name} | epoch={epoch} | step={global_step} | {epoch_recorder} | lowest value={lowest_value_rounded} (epoch {lowest_value_epoch} and step {lowest_value_step}) | remaining epochs for overtraining: g/total: {remaining_epochs_gen} d/total: {remaining_epochs_disc} | smoothed generator loss={smoothed_value_gen} | smoothed discriminator loss={smoothed_value_disc}",
|
515 |
+
"model_training_info_2": "{model_name} | epoch={epoch} | step={global_step} | {epoch_recorder} | lowest value={lowest_value_rounded} (epoch {lowest_value_epoch} and step {lowest_value_step})",
|
516 |
+
"model_training_info_3": "{model_name} | epoch={epoch} | step={global_step} | {epoch_recorder}",
|
517 |
+
"training_error": "An error occurred while training the model:",
|
518 |
+
"separator_info": "Initializing with output path: {output_dir}, output format: {output_format}",
|
519 |
+
"output_dir_is_none": "Output folder not specified. Using current working directory.",
|
520 |
+
">0or=1": "Normalization threshold must be greater than 0 and less than or equal to 1.",
|
521 |
+
"output_single": "Single root output requested; only one file ({output_single_stem}) will be written",
|
522 |
+
"step2": "The second step will be reversed using spectrogram instead of waveform. This may improve quality but is slightly slower.",
|
523 |
+
"other_than_zero": "Sample rate set to {sample_rate} but must be a nonzero integer.",
|
524 |
+
"too_high": "Sample rate set to {sample_rate}. Enter a less ambitious value.",
|
525 |
+
"sr_not_valid": "Sample rate must be a nonzero integer. Please provide a valid integer.",
|
526 |
+
"name_ver": "Version {name}",
|
527 |
+
"os": "Operating System",
|
528 |
+
"platform_info": "System: {system_info} Name: {node} Release: {release} Machine: {machine} Processor: {processor}",
|
529 |
+
"install_ffmpeg": "FFmpeg installed",
|
530 |
+
"none_ffmpeg": "FFmpeg is not installed. Please install FFmpeg to use this package.",
|
531 |
+
"install_onnx": "ONNX Runtime package {pu} installed with version",
|
532 |
+
"running_in_cpu": "Unable to configure hardware acceleration, running in CPU mode",
|
533 |
+
"running_in_cuda": "CUDA available in Torch, setting Torch device to CUDA",
|
534 |
+
"onnx_have": "ONNXruntime available {have}, enabling acceleration",
|
535 |
+
"onnx_not_have": "{have} not available in ONNXruntime; acceleration will NOT be enabled",
|
536 |
+
"python_not_install": "Python package: {package_name} is not installed",
|
537 |
+
"hash": "Calculating hash for model file {model_path}",
|
538 |
+
"ioerror": "IOError while seeking -10 MB or reading model file to compute hash: {e}",
|
539 |
+
"cancel_download": "File already exists at {output_path}, skipping download",
|
540 |
+
"download_model": "Downloading file from {url} to {output_path} with a timeout of 300 seconds",
|
541 |
+
"download_error": "Failed to download file from {url}, response code: {status_code}",
|
542 |
+
"vip_model": "Model: '{model_friendly_name}' is a premium model intended by Anjok07 only for paid subscriber access.",
|
543 |
+
"vip_print": "Hey there, if you haven't subscribed, please consider supporting UVR's developer, Anjok07, by subscribing here: https://patreon.com/uvr",
|
544 |
+
"search_model": "Searching for model {model_filename} in the list of supported models in the group",
|
545 |
+
"load_download_json": "Downloaded model list loaded",
|
546 |
+
"single_model": "Identified single model file: {model_friendly_name}",
|
547 |
+
"not_found_model": "Model not found in the UVR repository, attempting to download from the audio model separation repository...",
|
548 |
+
"single_model_path": "Returning path for single model file: {model_path}",
|
549 |
+
"find_model": "Input file name {model_filename} found in multi-file model: {model_friendly_name}",
|
550 |
+
"find_models": "Identified multi-file model: {model_friendly_name}, iterating through files to download",
|
551 |
+
"find_path": "Attempting to determine download PATH for config pair",
|
552 |
+
"not_found_model_warehouse": "Model not found in the UVR repository, attempting to download from the audio model separation repository...",
|
553 |
+
"yaml_warning": "The model name you specified, {model_filename}, is actually a model config file rather than a model file.",
|
554 |
+
"yaml_warning_2": "We found a model matching this config file: {config_key}, so we’ll use that model file for this run.",
|
555 |
+
"yaml_warning_3": "To avoid confusing/inconsistent behavior in the future, specify the actual model file name instead.",
|
556 |
+
"yaml_debug": "Config YAML model file not found in UVR repository, attempting to download from the audio model separation repository...",
|
557 |
+
"download_model_friendly": "All files downloaded for model {model_friendly_name}, returning original path {model_path}",
|
558 |
+
"not_found_model_2": "Model file {model_filename} not found in the supported files",
|
559 |
+
"load_yaml": "Loading model data from YAML at path {model_data_yaml_filepath}",
|
560 |
+
"load_yaml_2": "Model data loaded from YAML file: {model_data}",
|
561 |
+
"hash_md5": "Computing MD5 hash for model file to identify model parameters from UVR data...",
|
562 |
+
"model_hash": "Model {model_path} has hash {model_hash}",
|
563 |
+
"mdx_data": "MDX model data path set to {mdx_model_data_path}",
|
564 |
+
"load_mdx": "Loading MDX model parameters from UVR model data file...",
|
565 |
+
"model_not_support": "Unsupported model file: no parameters found for MD5 hash {model_hash} in UVR model data for MDX vault.",
|
566 |
+
"uvr_json": "Model data loaded from UVR JSON with hash {model_hash}: {model_data}",
|
567 |
+
"loading_model": "Loading model {model_filename}...",
|
568 |
+
"download_model_friendly_2": "Downloaded model, friendly name: {model_friendly_name}, Model path: {model_path}",
|
569 |
+
"model_type_not_support": "Unsupported model type: {model_type}",
|
570 |
+
"demucs_not_support_python<3.10": "Demucs models require Python version 3.10 or higher.",
|
571 |
+
"import_module": "Importing module for model type",
|
572 |
+
"initialization": "Initializing separator class for model type",
|
573 |
+
"loading_model_success": "Model loading completed.",
|
574 |
+
"loading_model_duration": "Model loading duration",
|
575 |
+
"starting_separator": "Starting separation process for audio file path",
|
576 |
+
"normalization": "Normalization threshold set to {normalization_threshold}, waveform will be scaled down to this maximum amplitude to prevent clipping.",
|
577 |
+
"loading_separator_model": "Downloading model {model_filename}...",
|
578 |
+
"separator_success_3": "Separation process completed.",
|
579 |
+
"separator_duration": "Separation duration",
|
580 |
+
"downloading_model": "Downloaded model, type: {model_type}, friendly name: {model_friendly_name}, Model path: {model_path}, Model data: {model_data_dict_size} items",
|
581 |
+
"not_found_model_signature": "Pre-trained model with signature {sig} not found.",
|
582 |
+
"repo_must_be_folder": "{repo} must exist and be a directory.",
|
583 |
+
"demucs_info": "Demucs parameters: Segment size = {segment_size}, Segment size active = {segments_enabled}",
|
584 |
+
"demucs_info_2": "Demucs parameters: Number of predictions = {shifts}, Overlap = {overlap}",
|
585 |
+
"start_demucs": "Demucs Separator initialization completed",
|
586 |
+
"start_separator": "Starting separation process...",
|
587 |
+
"prepare_mix": "Preparing mixture...",
|
588 |
+
"demix": "Mixture prepared for demixing. Shape: {shape}",
|
589 |
+
"cancel_mix": "Loading model for demixing...",
|
590 |
+
"model_review": "Model loaded and set to evaluation mode.",
|
591 |
+
"del_gpu_cache_after_demix": "Cleared model and GPU cache after demixing.",
|
592 |
+
"process_output_file": "Processing output file...",
|
593 |
+
"source_length": "Processing source array, source length is {source_length}",
|
594 |
+
"process_ver": "Processing source version...",
|
595 |
+
"set_map": "Set source map to {part} parts...",
|
596 |
+
"process_all_part": "Processing for all root parts...",
|
597 |
+
"skip_part": "Skipping root part {stem_name} as out_single_stem is set to {output_single_stem}...",
|
598 |
+
"starting_demix_demucs": "Starting the demix process in demix_demucs...",
|
599 |
+
"model_infer": "Running model inference...",
|
600 |
+
"del_all_but_one": "Duplicate pre-trained models detected for signature {xp_sig}. Please delete all but one.",
|
601 |
+
"name_not_pretrained": "{name} is not a pre-trained model or a model bundle.",
|
602 |
+
"invalid_checksum": "Invalid checksum for file {path}, expected {checksum} but got {actual_checksum}",
|
603 |
+
"mdx_info": "MDX parameters: Batch size = {batch_size}, Segment size = {segment_size}",
|
604 |
+
"mdx_info_2": "MDX parameters: Overlap = {overlap}, Hop_length = {hop_length}, Denoising enabled = {enable_denoise}",
|
605 |
+
"mdx_info_3": "MDX parameters",
|
606 |
+
"load_model_onnx": "Loading ONNX model for inference...",
|
607 |
+
"load_model_onnx_success": "Successfully loaded model using ONNXruntime inference session.",
|
608 |
+
"onnx_to_pytorch": "Model converted from ONNX to PyTorch due to mismatched segment size with dim_t, processing may be slower.",
|
609 |
+
"stft": "Inverse STFT applied. Returning result with shape",
|
610 |
+
"no_denoise": "Model running on spectrum without denoising.",
|
611 |
+
"mix": "Preparing mix for input audio file {audio_file_path}...",
|
612 |
+
"normalization_demix": "Normalizing mix prior to demixing...",
|
613 |
+
"mix_success": "Mix preparation completed.",
|
614 |
+
"primary_source": "Normalizing primary source...",
|
615 |
+
"secondary_source": "Producing secondary source: Mixing in compatible mode",
|
616 |
+
"invert_using_spec": "Inverting secondary stem using spectrum when invert_USE_spec is set to True",
|
617 |
+
"invert_using_spec_2": "Inverting secondary stem by subtracting transformed stem from the initial transformed mix",
|
618 |
+
"enable_denoise": "Model running on both positive and negative spectrums for denoising.",
|
619 |
+
"is_match_mix": "is_match_mix: Predicted spectrum obtained directly from STFT output.",
|
620 |
+
"save_secondary_stem_output_path": "Saving secondary stem {stem_name} to {stem_output_path}...",
|
621 |
+
"starting_model": "Initializing model settings...",
|
622 |
+
"input_info": "Model input parameters",
|
623 |
+
"model_settings": "Model settings",
|
624 |
+
"initialize_mix": "Initializing mix with is_ckpt = {is_ckpt}. Initial mix shape: {shape}",
|
625 |
+
"!=2": "Expected 2-channel audio signal but got {shape} channels",
|
626 |
+
"process_check": "Processing in checkpoint mode...",
|
627 |
+
"stft_2": "STFT applied to mix. Spectrum shape: {shape}",
|
628 |
+
"cache": "Computed padding",
|
629 |
+
"shape": "Mix shape after padding: {shape}, Number of parts: {num_chunks}",
|
630 |
+
"process_no_check": "Processing in no-checkpoint mode...",
|
631 |
+
"n_sample_or_pad": "Number of samples: {n_sample}, Computed padding: {pad}",
|
632 |
+
"shape_2": "Mix shape after padding",
|
633 |
+
"process_part": "Processed part {mix_waves}: Start {i}, End {ii}",
|
634 |
+
"mix_waves_to_tensor": "Converted mix_waves to tensor. Tensor shape: {shape}",
|
635 |
+
"mix_match": "Mix mode Match; applying compensation factor.",
|
636 |
+
"tar_waves": "tar_waves. Shape",
|
637 |
+
"normalization_2": "Normalizing result by dividing it by divisor.",
|
638 |
+
"mix_wave": "Processing mix_wave batch",
|
639 |
+
"mix_or_batch": "Mix parts into batches. Number of batches",
|
640 |
+
"demix_is_match_mix": "Starting demix process with is_match_mix,",
|
641 |
+
"mix_shape": "Root mix parts stored. Shape",
|
642 |
+
"chunk_size_or_overlap": "Chunk size for compatible mixing: {chunk_size}, Overlap: {overlap}",
|
643 |
+
"chunk_size_or_overlap_standard": "Standard chunk size: {chunk_size}, Overlap: {overlap}",
|
644 |
+
"calc_size": "Generated size calculated",
|
645 |
+
"window": "Window applied to this segment.",
|
646 |
+
"process_part_2": "Processing segment {total}/{total_chunks}: Start {start}, End {end}",
|
647 |
+
"all_process_part": "Total segments to process",
|
648 |
+
"step_or_overlap": "Step size to process parts: {step} with overlap set to {overlap}.",
|
649 |
+
"mix_cache": "Mix prepared with padding. Mix shape",
|
650 |
+
"dims": "Cannot use sin/cos position encoding with odd dimensions (dim={dims})",
|
651 |
+
"activation": "activation must be relu/gelu, not {activation}",
|
652 |
+
"length_or_training_length": "Provided length {length} exceeds training duration {training_length}",
|
653 |
+
"type_not_valid": "Invalid type for",
|
654 |
+
"del_parameter": "Removing non-existent parameter ",
|
655 |
+
"info": "Common parameters: Model name = {model_name}, Model path = {model_path}",
|
656 |
+
"info_2": "Common parameters: Output path = {output_dir}, Output format = {output_format}",
|
657 |
+
"info_3": "Common parameters: Normalization threshold = {normalization_threshold}",
|
658 |
+
"info_4": "Common parameters: Denoising enabled = {enable_denoise}, Single stem output = {output_single_stem}",
|
659 |
+
"info_5": "Common parameters: Inversion using specs = {invert_using_spec}, Sample rate = {sample_rate}",
|
660 |
+
"info_6": "Common parameters: Primary root name = {primary_stem_name}, Secondary root name = {secondary_stem_name}",
|
661 |
+
"info_7": "Common parameters: Karaoke mode = {is_karaoke}, BV model = {is_bv_model}, BV model rebalancing = {bv_model_rebalance}",
|
662 |
+
"success_process": "Completed processing root {stem_name} and writing audio...",
|
663 |
+
"load_audio": "Loading audio from file",
|
664 |
+
"load_audio_success": "Audio loaded. Sample rate: {sr}, Audio shape: {shape}",
|
665 |
+
"convert_mix": "Converting provided mix array.",
|
666 |
+
"convert_shape": "Converted mix shape: {shape}",
|
667 |
+
"audio_not_valid": "Audio file {audio_path} is empty or invalid",
|
668 |
+
"audio_valid": "Audio file is valid and contains data.",
|
669 |
+
"mix_single": "Mix is mono. Converting to stereo.",
|
670 |
+
"convert_mix_audio": "Converted to stereo mix.",
|
671 |
+
"mix_success_2": "Mix preparation completed.",
|
672 |
+
"duration": "Audio duration is {duration_hours} hours ({duration_seconds} seconds).",
|
673 |
+
"write": "Using {name} to write.",
|
674 |
+
"write_audio": "Writing {name} with root path:",
|
675 |
+
"original_not_valid": "Warning: Original source array is nearly silent or empty.",
|
676 |
+
"shape_audio": "Audio data shape before processing",
|
677 |
+
"convert_data": "Data type before conversion",
|
678 |
+
"original_source_to_int16": "Converted original_source to int16.",
|
679 |
+
"shape_audio_2": "Interleaved audio data shape",
|
680 |
+
"create_audiosegment": "Successfully created AudioSegment.",
|
681 |
+
"create_audiosegment_error": "Specific error while creating AudioSegment",
|
682 |
+
"export_error": "Error exporting audio file",
|
683 |
+
"export_success": "Successfully exported audio file to",
|
684 |
+
"clean": "Running garbage collection...",
|
685 |
+
"clean_cache": "Clearing {name} cache...",
|
686 |
+
"del_path": "Deleting path, source, and root of input audio file...",
|
687 |
+
"not_success": "Process was not successful: ",
|
688 |
+
"resample_error": "Error during resampling",
|
689 |
+
"shapes": "Shapes",
|
690 |
+
"wav_resolution": "Resolution type",
|
691 |
+
"warnings": "Warning: Extremely aggressive values detected",
|
692 |
+
"warnings_2": "Warning: NaN or infinite values detected in wave input. Shape",
|
693 |
+
"process_file": "Processing file... \n",
|
694 |
+
"save_instruments": "Saving reverse track...",
|
695 |
+
"assert": "Audio files must have the same shape - Mix: {mixshape}, Inst: {instrumentalshape}",
|
696 |
+
"rubberband": "Rubberband CLI cannot be executed. Please ensure Rubberband-CLI is installed.",
|
697 |
+
"rate": "Rate must be strictly positive",
|
698 |
+
"gdown_warning": "You provided a Google Drive link that is not a direct file download link. Consider using the `--fuzzy` option or the following URL: https://drive.google.com/uc?id={file_id}",
|
699 |
+
"gdown_error": "Could not retrieve the public link for the file. You may need to change its permissions to 'Anyone with the link' or there may already be excessive access permissions.",
|
700 |
+
"gdown_error_2": "Failed to retrieve file path:\n\n{error}\n\nYou can still access the file via a browser:\n\n\t{url_origin}\n\nbut Gdown cannot. Please check your connection and permissions.",
|
701 |
+
"temps": "Several temporary files exist for continuation:",
|
702 |
+
"del_all_temps": "Please delete all except one to continue downloading.",
|
703 |
+
"continue": "Continue:",
|
704 |
+
"to": "To:",
|
705 |
+
"gdown_value_error": "A path or ID must be specified",
|
706 |
+
"missing_url": "URL is missing",
|
707 |
+
"mac_not_match": "MAC does not match",
|
708 |
+
"file_not_access": "File is not accessible",
|
709 |
+
"int_resp==-3": "Request failed, retrying"
|
710 |
}
|