Shreyas094 commited on
Commit
c05359d
·
verified ·
1 Parent(s): f404999

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +55 -26
app.py CHANGED
@@ -475,7 +475,25 @@ def create_web_search_vectors(search_results):
475
 
476
  return FAISS.from_documents(documents, embed)
477
 
478
- def get_response_with_search(query, model, num_calls=3, temperature=0.2):
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
479
  search_results = duckduckgo_search(query)
480
  web_search_database = create_web_search_vectors(search_results)
481
 
@@ -483,36 +501,47 @@ def get_response_with_search(query, model, num_calls=3, temperature=0.2):
483
  yield "No web search results available. Please try again.", ""
484
  return
485
 
486
- retriever = web_search_database.as_retriever(search_kwargs={"k": 5})
487
  relevant_docs = retriever.get_relevant_documents(query)
488
 
489
- context = "\n".join([doc.page_content for doc in relevant_docs])
490
 
491
- prompt = f"""Using the following context from web search results:
492
  {context}
493
- Write a detailed and complete research document that fulfills the following user request: '{query}'
494
- After writing the document, please provide a list of sources used in your response."""
495
 
496
- if model == "@cf/meta/llama-3.1-8b-instruct":
497
- # Use Cloudflare API
498
- for response in get_response_from_cloudflare(prompt="", context=context, query=query, num_calls=num_calls, temperature=temperature, search_type="web"):
499
- yield response, "" # Yield streaming response without sources
500
- else:
501
- # Use Hugging Face API
502
- client = InferenceClient(model, token=huggingface_token)
503
-
504
- main_content = ""
505
- for i in range(num_calls):
506
- for message in client.chat_completion(
507
- messages=[{"role": "user", "content": prompt}],
508
- max_tokens=10000,
509
- temperature=temperature,
510
- stream=True,
511
- ):
512
- if message.choices and message.choices[0].delta and message.choices[0].delta.content:
513
- chunk = message.choices[0].delta.content
514
- main_content += chunk
515
- yield main_content, "" # Yield partial main content without sources
 
 
 
 
 
 
 
 
 
 
 
516
 
517
 
518
  INSTRUCTION_PROMPTS = {
 
475
 
476
  return FAISS.from_documents(documents, embed)
477
 
478
+ def critique_response(response, context, query):
479
+ critique_prompt = f"""Given the following response, original context, and user query, identify any statements that might be inaccurate, unsupported by the context, or irrelevant to the query. Be specific about which parts may be hallucinations or extrapolations beyond the given information.
480
+
481
+ User Query: {query}
482
+
483
+ Response:
484
+ {response}
485
+
486
+ Original Context:
487
+ {context}
488
+
489
+ Critique:"""
490
+
491
+ client = InferenceClient(model, token=huggingface_token)
492
+ critique = client.text_generation(critique_prompt, max_new_tokens=500, temperature=0.2)
493
+
494
+ return critique
495
+
496
+ def get_response_with_search(query, model, num_calls=3, temperature=0.1):
497
  search_results = duckduckgo_search(query)
498
  web_search_database = create_web_search_vectors(search_results)
499
 
 
501
  yield "No web search results available. Please try again.", ""
502
  return
503
 
504
+ retriever = web_search_database.as_retriever(search_kwargs={"k": 10})
505
  relevant_docs = retriever.get_relevant_documents(query)
506
 
507
+ context = "\n".join([doc.page_content for doc in relevant_docs[:5]])
508
 
509
+ initial_prompt = f"""Using the following context from web search results:
510
  {context}
511
+ Write a detailed and complete response that answers the following user query: '{query}'
512
+ Stick closely to the information provided in the context and avoid making unsupported claims."""
513
 
514
+ client = InferenceClient(model, token=huggingface_token)
515
+
516
+ # Generate initial response
517
+ initial_response = client.text_generation(initial_prompt, max_new_tokens=1000, temperature=temperature)
518
+
519
+ # Generate critique
520
+ critique = critique_response(initial_response, context, query)
521
+
522
+ final_prompt = f"""Given the following initial response, context, critique, and original query, provide a revised response that addresses the identified issues and sticks closely to the information provided in the context while fully answering the user's query.
523
+
524
+ User Query: {query}
525
+
526
+ Initial Response:
527
+ {initial_response}
528
+
529
+ Context:
530
+ {context}
531
+
532
+ Critique:
533
+ {critique}
534
+
535
+ Revised Response:"""
536
+
537
+ # Generate final response
538
+ for chunk in client.text_generation(final_prompt, max_new_tokens=1500, temperature=temperature, stream=True):
539
+ yield chunk, ""
540
+
541
+ # Add a disclaimer
542
+ disclaimer = ("\nNote: This response was generated by an AI model based on web search results. "
543
+ "While efforts have been made to ensure accuracy, please verify important information from authoritative sources.")
544
+ yield disclaimer, ""
545
 
546
 
547
  INSTRUCTION_PROMPTS = {