Spaces:
Sleeping
Sleeping
Update main.py
Browse files
main.py
CHANGED
@@ -37,24 +37,17 @@ if not TOGETHER_AI_API:
|
|
37 |
try:
|
38 |
embeddings = HuggingFaceEmbeddings(
|
39 |
model_name="nomic-ai/nomic-embed-text-v1",
|
40 |
-
model_kwargs={"trust_remote_code": True},
|
41 |
)
|
42 |
except Exception as e:
|
43 |
logger.error(f"Error loading embeddings: {e}")
|
44 |
raise RuntimeError("Failed to initialize embeddings.")
|
45 |
|
46 |
# Load FAISS vectorstore
|
47 |
-
|
48 |
-
|
49 |
-
|
50 |
-
|
51 |
-
logger.error(f"Error loading FAISS vectorstore: {e}")
|
52 |
-
loader = DirectoryLoader('./data')
|
53 |
-
text_splitter = RecursiveCharacterTextSplitter(chunk_size=500, chunk_overlap=50)
|
54 |
-
documents = text_splitter.split_documents(loader.load())
|
55 |
-
db = FAISS.from_documents(documents, embeddings)
|
56 |
-
db.save_local("ipc_vector_db")
|
57 |
-
db_retriever = db.as_retriever(search_type="similarity", search_kwargs={"k": 5, "score_threshold": 0.8})
|
58 |
|
59 |
# Define prompt template
|
60 |
prompt_template = """
|
@@ -71,8 +64,8 @@ prompt = PromptTemplate(template=prompt_template, input_variables=["context", "q
|
|
71 |
try:
|
72 |
llm = Together(
|
73 |
model="mistralai/Mistral-7B-Instruct-v0.2",
|
74 |
-
temperature=0.
|
75 |
-
max_tokens=
|
76 |
together_api_key=TOGETHER_AI_API,
|
77 |
)
|
78 |
except Exception as e:
|
|
|
37 |
try:
|
38 |
embeddings = HuggingFaceEmbeddings(
|
39 |
model_name="nomic-ai/nomic-embed-text-v1",
|
40 |
+
model_kwargs={"trust_remote_code": True,"revision":"289f532e14dbbbd5a04753fa58739e9ba766f3c7"},
|
41 |
)
|
42 |
except Exception as e:
|
43 |
logger.error(f"Error loading embeddings: {e}")
|
44 |
raise RuntimeError("Failed to initialize embeddings.")
|
45 |
|
46 |
# Load FAISS vectorstore
|
47 |
+
|
48 |
+
db = FAISS.load_local("ipc_vector_db", embeddings, allow_dangerous_deserialization=True)
|
49 |
+
db_retriever = db.as_retriever(search_type="similarity", search_kwargs={"k": 2, "max-length":512})
|
50 |
+
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
51 |
|
52 |
# Define prompt template
|
53 |
prompt_template = """
|
|
|
64 |
try:
|
65 |
llm = Together(
|
66 |
model="mistralai/Mistral-7B-Instruct-v0.2",
|
67 |
+
temperature=0.5,
|
68 |
+
max_tokens=1024,
|
69 |
together_api_key=TOGETHER_AI_API,
|
70 |
)
|
71 |
except Exception as e:
|