Update app.py
Browse files
app.py
CHANGED
@@ -169,10 +169,10 @@ class LlamaCppModel(Model):
|
|
169 |
model_path=model_path,
|
170 |
flash_attn=False,
|
171 |
n_gpu_layers=0,
|
172 |
-
n_batch=
|
173 |
n_ctx=n_ctx,
|
174 |
n_threads=2,
|
175 |
-
n_threads_batch=
|
176 |
)
|
177 |
|
178 |
elif repo_id and filename:
|
@@ -263,7 +263,7 @@ class LlamaCppModel(Model):
|
|
263 |
llm_sampling_settings=settings,
|
264 |
chat_history=messages,
|
265 |
returns_streaming_generator=False,
|
266 |
-
print_output=
|
267 |
|
268 |
)
|
269 |
|
@@ -298,7 +298,6 @@ Reference
|
|
298 |
- [Dataset(m-ric/huggingface_doc)](https://huggingface.co/datasets/m-ric/huggingface_doc)
|
299 |
|
300 |
"""
|
301 |
-
#Tool not support
|
302 |
agent = CodeAgent(prompt_templates =yaml.safe_load(prompt),model=model, tools=[retriever_tool],max_steps=1,verbosity_level=0,name="AGENT",description=description)
|
303 |
|
304 |
demo = GradioUI(agent)
|
|
|
169 |
model_path=model_path,
|
170 |
flash_attn=False,
|
171 |
n_gpu_layers=0,
|
172 |
+
#n_batch=1024,
|
173 |
n_ctx=n_ctx,
|
174 |
n_threads=2,
|
175 |
+
n_threads_batch=2,verbose=False
|
176 |
)
|
177 |
|
178 |
elif repo_id and filename:
|
|
|
263 |
llm_sampling_settings=settings,
|
264 |
chat_history=messages,
|
265 |
returns_streaming_generator=False,
|
266 |
+
print_output=False,
|
267 |
|
268 |
)
|
269 |
|
|
|
298 |
- [Dataset(m-ric/huggingface_doc)](https://huggingface.co/datasets/m-ric/huggingface_doc)
|
299 |
|
300 |
"""
|
|
|
301 |
agent = CodeAgent(prompt_templates =yaml.safe_load(prompt),model=model, tools=[retriever_tool],max_steps=1,verbosity_level=0,name="AGENT",description=description)
|
302 |
|
303 |
demo = GradioUI(agent)
|