Update myagent.py
Browse files- myagent.py +4 -1
myagent.py
CHANGED
@@ -60,7 +60,7 @@ class LocalLlamaModel:
|
|
60 |
self.tokenizer = tokenizer
|
61 |
self.device = 'cpu'
|
62 |
|
63 |
-
def generate(self, prompt: str, max_new_tokens=512, **kwargs):
|
64 |
try:
|
65 |
# Generate answer using the provided prompt - following the recommended pattern
|
66 |
# input_ids = self.tokenizer.apply_chat_template(
|
@@ -69,6 +69,9 @@ class LocalLlamaModel:
|
|
69 |
# return_tensors="pt",
|
70 |
# tokenize=True,
|
71 |
# ).to(self.model.device)
|
|
|
|
|
|
|
72 |
|
73 |
inputs = self.tokenizer(prompt, return_tensors="pt").to(self.model.device)
|
74 |
|
|
|
60 |
self.tokenizer = tokenizer
|
61 |
self.device = 'cpu'
|
62 |
|
63 |
+
def generate(self, prompt: str, max_new_tokens=512*5, **kwargs):
|
64 |
try:
|
65 |
# Generate answer using the provided prompt - following the recommended pattern
|
66 |
# input_ids = self.tokenizer.apply_chat_template(
|
|
|
69 |
# return_tensors="pt",
|
70 |
# tokenize=True,
|
71 |
# ).to(self.model.device)
|
72 |
+
|
73 |
+
print("Prompt: ", prompt)
|
74 |
+
print("Prompt type: ", type(prompt))
|
75 |
|
76 |
inputs = self.tokenizer(prompt, return_tensors="pt").to(self.model.device)
|
77 |
|