Shreyas094 commited on
Commit
b5da50d
·
verified ·
1 Parent(s): 6048ff3

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -3
app.py CHANGED
@@ -525,10 +525,9 @@ def get_response_from_llama(query, model, selected_docs, file_type, num_calls=1,
525
  # Generate content with streaming enabled
526
  for message in client.chat_completion(
527
  messages=messages,
528
- max_new_tokens=20000,
529
  temperature=temperature,
530
  stream=True,
531
- top_p=0.8,
532
  ):
533
  if message.choices and message.choices[0].delta and message.choices[0].delta.content:
534
  chunk = message.choices[0].delta.content
@@ -776,7 +775,7 @@ def get_response_from_pdf(query, model, selected_docs, num_calls=3, temperature=
776
  logging.info(f"API call {i+1}/{num_calls}")
777
  for message in client.chat_completion(
778
  messages=messages,
779
- max_new_tokens=20000,
780
  temperature=temperature,
781
  stream=True,
782
  top_p=0.8,
 
525
  # Generate content with streaming enabled
526
  for message in client.chat_completion(
527
  messages=messages,
528
+ max_tokens=20000,
529
  temperature=temperature,
530
  stream=True,
 
531
  ):
532
  if message.choices and message.choices[0].delta and message.choices[0].delta.content:
533
  chunk = message.choices[0].delta.content
 
775
  logging.info(f"API call {i+1}/{num_calls}")
776
  for message in client.chat_completion(
777
  messages=messages,
778
+ max_tokens=20000,
779
  temperature=temperature,
780
  stream=True,
781
  top_p=0.8,