Update app.py
Browse files
app.py
CHANGED
@@ -121,16 +121,30 @@ retriever = db.as_retriever(
|
|
121 |
|
122 |
|
123 |
# Set up the LLM
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
124 |
llm = ChatOpenAI(
|
125 |
-
base_url="https://
|
126 |
temperature=0,
|
127 |
api_key=HF_TOKEN,
|
128 |
-
model="
|
129 |
max_tokens=None,
|
130 |
timeout=None
|
131 |
|
132 |
)
|
133 |
|
|
|
|
|
|
|
134 |
# quantization_config = BitsAndBytesConfig(
|
135 |
# load_in_4bit=True,
|
136 |
# bnb_4bit_compute_dtype=torch.bfloat16,
|
|
|
121 |
|
122 |
|
123 |
# Set up the LLM
|
124 |
+
# llm = ChatOpenAI(
|
125 |
+
# base_url="https://api-inference.huggingface.co/v1/",
|
126 |
+
# temperature=0,
|
127 |
+
# api_key=HF_TOKEN,
|
128 |
+
# model="mistralai/Mistral-Nemo-Instruct-2407",
|
129 |
+
# max_tokens=None,
|
130 |
+
# timeout=None
|
131 |
+
|
132 |
+
# )
|
133 |
+
|
134 |
+
|
135 |
llm = ChatOpenAI(
|
136 |
+
base_url="https://openrouter.ai/api/v1",
|
137 |
temperature=0,
|
138 |
api_key=HF_TOKEN,
|
139 |
+
model="google/gemini-2.0-flash-exp:free",
|
140 |
max_tokens=None,
|
141 |
timeout=None
|
142 |
|
143 |
)
|
144 |
|
145 |
+
|
146 |
+
|
147 |
+
|
148 |
# quantization_config = BitsAndBytesConfig(
|
149 |
# load_in_4bit=True,
|
150 |
# bnb_4bit_compute_dtype=torch.bfloat16,
|