Update app.py
Browse files
app.py
CHANGED
@@ -36,7 +36,7 @@ def get_pdf_processed(pdf_docs):
|
|
36 |
return text
|
37 |
|
38 |
def llm_model():
|
39 |
-
llm = ChatGroq(model="mixtral-8x7b-32768"
|
40 |
prompt = ChatPromptTemplate.from_template(
|
41 |
"""
|
42 |
Answer the question based on the provided context only.
|
@@ -60,7 +60,7 @@ def llm_model():
|
|
60 |
print("Response time :", time.process_time()-start)
|
61 |
st.write(response['answer'])
|
62 |
|
63 |
-
st.session_state.embeddings =GoogleGenerativeAIEmbeddings(model = 'models/embedding-001'
|
64 |
st.session_state.text_splitter = RecursiveCharacterTextSplitter(chunk_size =1000, chunk_overlap= 200)
|
65 |
|
66 |
if option:
|
|
|
36 |
return text
|
37 |
|
38 |
def llm_model():
|
39 |
+
llm = ChatGroq(model="mixtral-8x7b-32768")
|
40 |
prompt = ChatPromptTemplate.from_template(
|
41 |
"""
|
42 |
Answer the question based on the provided context only.
|
|
|
60 |
print("Response time :", time.process_time()-start)
|
61 |
st.write(response['answer'])
|
62 |
|
63 |
+
st.session_state.embeddings =GoogleGenerativeAIEmbeddings(model = 'models/embedding-001')
|
64 |
st.session_state.text_splitter = RecursiveCharacterTextSplitter(chunk_size =1000, chunk_overlap= 200)
|
65 |
|
66 |
if option:
|