Spaces:
Runtime error
Runtime error
Update pages/jarvis.py
Browse files- pages/jarvis.py +17 -20
pages/jarvis.py
CHANGED
|
@@ -77,32 +77,29 @@ def main():
|
|
| 77 |
chunk_size = st.slider("Chunk size", min_value=100, max_value=1000, value=600, step=20)
|
| 78 |
chunk_overlap = st.slider("Chunk overlap", min_value=10, max_value=200, value=40, step=10)
|
| 79 |
|
| 80 |
-
|
| 81 |
-
selected_tab = st.radio("Select Tab", tabs)
|
| 82 |
|
| 83 |
-
if
|
| 84 |
-
|
| 85 |
-
|
| 86 |
-
|
| 87 |
-
|
| 88 |
-
|
| 89 |
|
| 90 |
-
|
| 91 |
-
|
| 92 |
-
|
| 93 |
-
|
| 94 |
-
top_k = st.slider("Top-K Samples", min_value=1, max_value=10, value=3, step=1)
|
| 95 |
|
| 96 |
-
|
|
|
|
|
|
|
| 97 |
|
| 98 |
-
|
| 99 |
-
qa_chain = initialize_llmchain(llm_model, temperature, max_tokens, top_k, st.session_state["vector_db"])
|
| 100 |
-
|
| 101 |
-
elif selected_tab == "Chatbot":
|
| 102 |
-
if "qa_chain" in st.session_state:
|
| 103 |
message = st.text_input("Type your message")
|
| 104 |
if st.button("Submit"):
|
| 105 |
-
|
|
|
|
| 106 |
st.write("Assistant:", response["answer"])
|
| 107 |
|
| 108 |
if __name__ == "__main__":
|
|
|
|
| 77 |
chunk_size = st.slider("Chunk size", min_value=100, max_value=1000, value=600, step=20)
|
| 78 |
chunk_overlap = st.slider("Chunk overlap", min_value=10, max_value=200, value=40, step=10)
|
| 79 |
|
| 80 |
+
list_file_path = [file.name for file in uploaded_files]
|
|
|
|
| 81 |
|
| 82 |
+
if st.button("Generate Vector Database"):
|
| 83 |
+
st.text("Loading documents...")
|
| 84 |
+
doc_splits = load_doc(list_file_path, chunk_size, chunk_overlap)
|
| 85 |
+
st.text("Creating vector database...")
|
| 86 |
+
collection_name = create_collection_name(list_file_path[0])
|
| 87 |
+
vector_db = create_db(doc_splits, collection_name)
|
| 88 |
|
| 89 |
+
llm_model = st.selectbox("Choose LLM Model", list_llm)
|
| 90 |
+
temperature = st.slider("Temperature", min_value=0.01, max_value=1.0, value=0.7, step=0.1)
|
| 91 |
+
max_tokens = st.slider("Max Tokens", min_value=224, max_value=4096, value=1024, step=32)
|
| 92 |
+
top_k = st.slider("Top-K Samples", min_value=1, max_value=10, value=3, step=1)
|
|
|
|
| 93 |
|
| 94 |
+
if st.button("Initialize QA Chain"):
|
| 95 |
+
st.text("Initializing QA chain...")
|
| 96 |
+
qa_chain = initialize_llmchain(llm_model, temperature, max_tokens, top_k, vector_db)
|
| 97 |
|
| 98 |
+
st.header("Chatbot")
|
|
|
|
|
|
|
|
|
|
|
|
|
| 99 |
message = st.text_input("Type your message")
|
| 100 |
if st.button("Submit"):
|
| 101 |
+
st.text("Generating response...")
|
| 102 |
+
response = qa_chain({"question": message, "chat_history": []})
|
| 103 |
st.write("Assistant:", response["answer"])
|
| 104 |
|
| 105 |
if __name__ == "__main__":
|