Synced repo using 'sync_with_huggingface' Github Action
Browse files- Dockerfile +1 -0
- app.py +1 -1
- scrape_chroma.py +1 -1
Dockerfile
CHANGED
@@ -11,6 +11,7 @@ RUN mkdir -p /data/hf_cache/transformers /data/hf_cache/datasets /data/hf_cache/
|
|
11 |
|
12 |
# Ensure /app is writable
|
13 |
RUN chmod -R 777 /app
|
|
|
14 |
|
15 |
COPY requirements.txt .
|
16 |
RUN pip install --no-cache-dir -r requirements.txt
|
|
|
11 |
|
12 |
# Ensure /app is writable
|
13 |
RUN chmod -R 777 /app
|
14 |
+
RUN mkdir -p /app/chroma_db && chmod -R 777 /app/chroma_db
|
15 |
|
16 |
COPY requirements.txt .
|
17 |
RUN pip install --no-cache-dir -r requirements.txt
|
app.py
CHANGED
@@ -34,7 +34,7 @@ def store_feedback_csv(claim, result, feedback, csv_file="feedback_log.csv"):
|
|
34 |
|
35 |
def initialize_services():
|
36 |
return FactChecker(
|
37 |
-
chroma_path="
|
38 |
collection_name="pib_titles",
|
39 |
groq_client=OpenAI(
|
40 |
api_key=os.getenv("GROQ_API_KEY"),
|
|
|
34 |
|
35 |
def initialize_services():
|
36 |
return FactChecker(
|
37 |
+
chroma_path="app/chroma_db",
|
38 |
collection_name="pib_titles",
|
39 |
groq_client=OpenAI(
|
40 |
api_key=os.getenv("GROQ_API_KEY"),
|
scrape_chroma.py
CHANGED
@@ -7,7 +7,7 @@ import gc
|
|
7 |
import csv
|
8 |
|
9 |
# === CONFIGURATION ===
|
10 |
-
CHROMA_PATH = "chroma_db"
|
11 |
COLLECTION_NAME = "pib_titles"
|
12 |
|
13 |
def save_titles_to_csv(titles, filename="pib_titles.csv"):
|
|
|
7 |
import csv
|
8 |
|
9 |
# === CONFIGURATION ===
|
10 |
+
CHROMA_PATH = "app/chroma_db"
|
11 |
COLLECTION_NAME = "pib_titles"
|
12 |
|
13 |
def save_titles_to_csv(titles, filename="pib_titles.csv"):
|