Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -61,8 +61,8 @@ def load_pdf(file_path):
|
|
61 |
return VectorStore
|
62 |
|
63 |
|
64 |
-
def load_chatbot():
|
65 |
-
return load_qa_chain(llm=OpenAI(temperature=0.
|
66 |
|
67 |
|
68 |
def display_chat_history(chat_history):
|
@@ -97,10 +97,18 @@ def main():
|
|
97 |
loading_message.text('Bot is thinking...')
|
98 |
|
99 |
VectorStore = load_pdf(pdf)
|
100 |
-
|
|
|
101 |
docs = VectorStore.similarity_search(query=query, k=3)
|
102 |
with get_openai_callback() as cb:
|
103 |
response = chain.run(input_documents=docs, question=query)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
104 |
|
105 |
st.session_state['chat_history'].append(("Bot", response, "new"))
|
106 |
|
@@ -121,5 +129,8 @@ def main():
|
|
121 |
# Mark all messages as old after displaying
|
122 |
st.session_state['chat_history'] = [(sender, msg, "old") for sender, msg, _ in st.session_state['chat_history']]
|
123 |
|
|
|
|
|
|
|
124 |
if __name__ == "__main__":
|
125 |
main()
|
|
|
61 |
return VectorStore
|
62 |
|
63 |
|
64 |
+
def load_chatbot(max_tokens=120):
|
65 |
+
return load_qa_chain(llm=OpenAI(temperature=0.5, max_tokens=max_tokens), chain_type="stuff")
|
66 |
|
67 |
|
68 |
def display_chat_history(chat_history):
|
|
|
97 |
loading_message.text('Bot is thinking...')
|
98 |
|
99 |
VectorStore = load_pdf(pdf)
|
100 |
+
max_tokens = 100 # Initial max tokens
|
101 |
+
chain = load_chatbot(max_tokens=max_tokens)
|
102 |
docs = VectorStore.similarity_search(query=query, k=3)
|
103 |
with get_openai_callback() as cb:
|
104 |
response = chain.run(input_documents=docs, question=query)
|
105 |
+
|
106 |
+
# Check if the response ends with a sentence-ending punctuation
|
107 |
+
while not response.strip().endswith(('.', '!', '?')) and max_tokens < MAX_TOKEN_LIMIT:
|
108 |
+
max_tokens += 50 # Increase the max_tokens limit
|
109 |
+
chain = load_chatbot(max_tokens=max_tokens)
|
110 |
+
additional_response = chain.run(input_documents=docs, question=query)
|
111 |
+
response += additional_response # Append the additional response to the original response
|
112 |
|
113 |
st.session_state['chat_history'].append(("Bot", response, "new"))
|
114 |
|
|
|
129 |
# Mark all messages as old after displaying
|
130 |
st.session_state['chat_history'] = [(sender, msg, "old") for sender, msg, _ in st.session_state['chat_history']]
|
131 |
|
132 |
+
# Define a maximum token limit to avoid infinite loops
|
133 |
+
MAX_TOKEN_LIMIT = 400
|
134 |
+
|
135 |
if __name__ == "__main__":
|
136 |
main()
|