Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -541,7 +541,7 @@ def get_response_from_llama(query, model, selected_docs, file_type, num_calls=1,
|
|
541 |
# Generate content with streaming enabled
|
542 |
for response in client.chat_completion(
|
543 |
messages=messages, # Pass messages in the required format
|
544 |
-
max_tokens=
|
545 |
temperature=temperature,
|
546 |
stream=True,
|
547 |
top_p=0.9,
|
|
|
541 |
# Generate content with streaming enabled
|
542 |
for response in client.chat_completion(
|
543 |
messages=messages, # Pass messages in the required format
|
544 |
+
max_tokens=3000, # Reduced to ensure we stay within token limits
|
545 |
temperature=temperature,
|
546 |
stream=True,
|
547 |
top_p=0.9,
|