Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -73,7 +73,7 @@ languages = {
|
|
73 |
# Global variables to hold component references
|
74 |
components = {}
|
75 |
dotenv.load_dotenv()
|
76 |
-
seamless_client = Client("https://facebook-seamless-m4t-v2-large.hf.space/--replicas/6w5sk/")
|
77 |
HuggingFace_Token = os.getenv("HuggingFace_Token")
|
78 |
hf_token = os.getenv("HuggingFace_Token")
|
79 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
@@ -322,7 +322,7 @@ def wrap_text(text, width=90):
|
|
322 |
wrapped_text = '\n'.join(wrapped_lines)
|
323 |
return wrapped_text
|
324 |
|
325 |
-
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen-1_8B-Chat", trust_remote_code=True)
|
326 |
model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen-1_8B-Chat", device_map="auto", trust_remote_code=True).eval()
|
327 |
|
328 |
class ChatBot:
|
|
|
73 |
# Global variables to hold component references
|
74 |
components = {}
|
75 |
dotenv.load_dotenv()
|
76 |
+
seamless_client = Client("https://facebook-seamless-m4t-v2-large.hf.space/--replicas/6w5sk/") #TruEra
|
77 |
HuggingFace_Token = os.getenv("HuggingFace_Token")
|
78 |
hf_token = os.getenv("HuggingFace_Token")
|
79 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
|
322 |
wrapped_text = '\n'.join(wrapped_lines)
|
323 |
return wrapped_text
|
324 |
|
325 |
+
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen-1_8B-Chat", trust_remote_code=True) #TruEra
|
326 |
model = AutoModelForCausalLM.from_pretrained("Qwen/Qwen-1_8B-Chat", device_map="auto", trust_remote_code=True).eval()
|
327 |
|
328 |
class ChatBot:
|