Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -143,8 +143,12 @@ def get_conversation_chain(vectorstore:FAISS) -> ConversationalRetrievalChain:
|
|
143 |
# #repo_id="TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF"
|
144 |
# model_kwargs={"temperature": 0.5, "max_length": 2096},#1048
|
145 |
#)
|
|
|
|
|
|
|
|
|
146 |
llm = HuggingFaceHub(
|
147 |
-
repo_id="
|
148 |
model_kwargs={"temperature": 0.5, "max_new_tokens": 1024, "max_length": 1048, "top_k": 3, "trust_remote_code": True, "torch_dtype": "auto"},
|
149 |
)
|
150 |
# llm = ChatOpenAI(temperature=0, model="gpt-3.5-turbo-0613")
|
|
|
143 |
# #repo_id="TheBloke/Mixtral-8x7B-Instruct-v0.1-GGUF"
|
144 |
# model_kwargs={"temperature": 0.5, "max_length": 2096},#1048
|
145 |
#)
|
146 |
+
#llm = HuggingFaceHub(
|
147 |
+
# repo_id="mistralai/Mixtral-8x7B-Instruct-v0.1",
|
148 |
+
# model_kwargs={"temperature": 0.5, "max_new_tokens": 1024, "max_length": 1048, "top_k": 3, "trust_remote_code": True, "torch_dtype": "auto"},
|
149 |
+
#)
|
150 |
llm = HuggingFaceHub(
|
151 |
+
repo_id="microsoft/Orca-2-13b",
|
152 |
model_kwargs={"temperature": 0.5, "max_new_tokens": 1024, "max_length": 1048, "top_k": 3, "trust_remote_code": True, "torch_dtype": "auto"},
|
153 |
)
|
154 |
# llm = ChatOpenAI(temperature=0, model="gpt-3.5-turbo-0613")
|