Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -525,10 +525,9 @@ def get_response_from_llama(query, model, selected_docs, file_type, num_calls=1,
|
|
525 |
# Generate content with streaming enabled
|
526 |
for message in client.chat_completion(
|
527 |
messages=messages,
|
528 |
-
|
529 |
temperature=temperature,
|
530 |
stream=True,
|
531 |
-
top_p=0.8,
|
532 |
):
|
533 |
if message.choices and message.choices[0].delta and message.choices[0].delta.content:
|
534 |
chunk = message.choices[0].delta.content
|
@@ -776,7 +775,7 @@ def get_response_from_pdf(query, model, selected_docs, num_calls=3, temperature=
|
|
776 |
logging.info(f"API call {i+1}/{num_calls}")
|
777 |
for message in client.chat_completion(
|
778 |
messages=messages,
|
779 |
-
|
780 |
temperature=temperature,
|
781 |
stream=True,
|
782 |
top_p=0.8,
|
|
|
525 |
# Generate content with streaming enabled
|
526 |
for message in client.chat_completion(
|
527 |
messages=messages,
|
528 |
+
max_tokens=20000,
|
529 |
temperature=temperature,
|
530 |
stream=True,
|
|
|
531 |
):
|
532 |
if message.choices and message.choices[0].delta and message.choices[0].delta.content:
|
533 |
chunk = message.choices[0].delta.content
|
|
|
775 |
logging.info(f"API call {i+1}/{num_calls}")
|
776 |
for message in client.chat_completion(
|
777 |
messages=messages,
|
778 |
+
max_tokens=20000,
|
779 |
temperature=temperature,
|
780 |
stream=True,
|
781 |
top_p=0.8,
|