Spaces:
Sleeping
Sleeping
Update LLM/Qwen.py
Browse files- LLM/Qwen.py +5 -3
LLM/Qwen.py
CHANGED
@@ -27,16 +27,18 @@ class Qwen:
|
|
27 |
|
28 |
return model, tokenizer
|
29 |
|
30 |
-
def generate(self, question, system_prompt=""):
|
|
|
|
|
31 |
if self.mode != 'api':
|
32 |
self.data["question"] = self.prompt + question
|
33 |
try:
|
34 |
response, self.history = self.model.chat(self.tokenizer, self.data["question"], history=self.history, system = system_prompt)
|
35 |
-
|
36 |
return response
|
37 |
except Exception as e:
|
38 |
print(e)
|
39 |
-
return "
|
40 |
else:
|
41 |
return self.predict_api(question)
|
42 |
def predict_api(self, question):
|
|
|
27 |
|
28 |
return model, tokenizer
|
29 |
|
30 |
+
def generate(self, question= "Hello Hello", system_prompt=""):
|
31 |
+
print("Qwen: ")
|
32 |
+
print(question)
|
33 |
if self.mode != 'api':
|
34 |
self.data["question"] = self.prompt + question
|
35 |
try:
|
36 |
response, self.history = self.model.chat(self.tokenizer, self.data["question"], history=self.history, system = system_prompt)
|
37 |
+
print(self.history)
|
38 |
return response
|
39 |
except Exception as e:
|
40 |
print(e)
|
41 |
+
return "\nSorry, your request has encountered an error. Please try again.\n"
|
42 |
else:
|
43 |
return self.predict_api(question)
|
44 |
def predict_api(self, question):
|