rcastriotta
publish
1a3fc6f
raw
history blame contribute delete
564 Bytes
import torch
from seamless_communication.inference import Translator
# Initialize a Translator object with a multitask model, vocoder on the GPU.
translator = Translator(
"seamlessM4T_v2_large", "vocoder_v2", torch.device("cuda:0"), torch.float16
)
def translate_text(text):
print("test")
# text_output, speech_output = translator.predict(
# input=text,
# task_str="T2ST",
# tgt_lang="spa",
# src_lang="eng",
# text_generation_opts=None,
# unit_generation_opts=None,
# )
# print(text_output)