Update app.py
Browse files
app.py
CHANGED
@@ -17,7 +17,7 @@ import time
|
|
17 |
import logging
|
18 |
|
19 |
# Setup logging
|
20 |
-
logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
|
21 |
logger = logging.getLogger(__name__)
|
22 |
|
23 |
# Persistent directory
|
@@ -34,10 +34,12 @@ for directory in [model_cache_dir, tool_cache_dir, file_cache_dir, report_dir, v
|
|
34 |
os.makedirs(directory, exist_ok=True)
|
35 |
|
36 |
os.environ["HF_HOME"] = model_cache_dir
|
37 |
-
os.environ["TRANSFORMERS_CACHE"] = model_cache_dir
|
38 |
os.environ["VLLM_CACHE_DIR"] = vllm_cache_dir
|
39 |
os.environ["TOKENIZERS_PARALLELISM"] = "false"
|
40 |
os.environ["CUDA_LAUNCH_BLOCKING"] = "1"
|
|
|
|
|
|
|
41 |
|
42 |
current_dir = os.path.dirname(os.path.abspath(__file__))
|
43 |
src_path = os.path.abspath(os.path.join(current_dir, "src"))
|
@@ -195,10 +197,9 @@ def init_agent():
|
|
195 |
tool_files_dict={"new_tool": target_tool_path},
|
196 |
force_finish=True,
|
197 |
enable_checker=True,
|
198 |
-
step_rag_num=1,
|
199 |
seed=100,
|
200 |
additional_default_tools=[],
|
201 |
-
num_engine_threads=1, # Limit VLLM threads for stability
|
202 |
)
|
203 |
agent.init_model()
|
204 |
log_system_usage("After Load")
|
@@ -230,8 +231,8 @@ def process_batch(agent, chunks: List[str], cache_path: str, prompt_template: st
|
|
230 |
message=prompt,
|
231 |
history=[],
|
232 |
temperature=0.2,
|
233 |
-
max_new_tokens=256,
|
234 |
-
max_token=1024,
|
235 |
call_agent=False,
|
236 |
conversation=[],
|
237 |
):
|
|
|
17 |
import logging
|
18 |
|
19 |
# Setup logging
|
20 |
+
logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s", filename="/home/user/clinical_oversight_analyzer.log")
|
21 |
logger = logging.getLogger(__name__)
|
22 |
|
23 |
# Persistent directory
|
|
|
34 |
os.makedirs(directory, exist_ok=True)
|
35 |
|
36 |
os.environ["HF_HOME"] = model_cache_dir
|
|
|
37 |
os.environ["VLLM_CACHE_DIR"] = vllm_cache_dir
|
38 |
os.environ["TOKENIZERS_PARALLELISM"] = "false"
|
39 |
os.environ["CUDA_LAUNCH_BLOCKING"] = "1"
|
40 |
+
# Remove TRANSFORMERS_CACHE to suppress warning
|
41 |
+
if "TRANSFORMERS_CACHE" in os.environ:
|
42 |
+
del os.environ["TRANSFORMERS_CACHE"]
|
43 |
|
44 |
current_dir = os.path.dirname(os.path.abspath(__file__))
|
45 |
src_path = os.path.abspath(os.path.join(current_dir, "src"))
|
|
|
197 |
tool_files_dict={"new_tool": target_tool_path},
|
198 |
force_finish=True,
|
199 |
enable_checker=True,
|
200 |
+
step_rag_num=1,
|
201 |
seed=100,
|
202 |
additional_default_tools=[],
|
|
|
203 |
)
|
204 |
agent.init_model()
|
205 |
log_system_usage("After Load")
|
|
|
231 |
message=prompt,
|
232 |
history=[],
|
233 |
temperature=0.2,
|
234 |
+
max_new_tokens=256,
|
235 |
+
max_token=1024,
|
236 |
call_agent=False,
|
237 |
conversation=[],
|
238 |
):
|