Ali2206 commited on
Commit
6358a36
·
verified ·
1 Parent(s): 28928c8

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -6
app.py CHANGED
@@ -17,7 +17,7 @@ import time
17
  import logging
18
 
19
  # Setup logging
20
- logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s")
21
  logger = logging.getLogger(__name__)
22
 
23
  # Persistent directory
@@ -34,10 +34,12 @@ for directory in [model_cache_dir, tool_cache_dir, file_cache_dir, report_dir, v
34
  os.makedirs(directory, exist_ok=True)
35
 
36
  os.environ["HF_HOME"] = model_cache_dir
37
- os.environ["TRANSFORMERS_CACHE"] = model_cache_dir
38
  os.environ["VLLM_CACHE_DIR"] = vllm_cache_dir
39
  os.environ["TOKENIZERS_PARALLELISM"] = "false"
40
  os.environ["CUDA_LAUNCH_BLOCKING"] = "1"
 
 
 
41
 
42
  current_dir = os.path.dirname(os.path.abspath(__file__))
43
  src_path = os.path.abspath(os.path.join(current_dir, "src"))
@@ -195,10 +197,9 @@ def init_agent():
195
  tool_files_dict={"new_tool": target_tool_path},
196
  force_finish=True,
197
  enable_checker=True,
198
- step_rag_num=1, # Reduced for speed
199
  seed=100,
200
  additional_default_tools=[],
201
- num_engine_threads=1, # Limit VLLM threads for stability
202
  )
203
  agent.init_model()
204
  log_system_usage("After Load")
@@ -230,8 +231,8 @@ def process_batch(agent, chunks: List[str], cache_path: str, prompt_template: st
230
  message=prompt,
231
  history=[],
232
  temperature=0.2,
233
- max_new_tokens=256, # Reduced for speed
234
- max_token=1024, # Reduced for speed
235
  call_agent=False,
236
  conversation=[],
237
  ):
 
17
  import logging
18
 
19
  # Setup logging
20
+ logging.basicConfig(level=logging.INFO, format="%(asctime)s - %(levelname)s - %(message)s", filename="/home/user/clinical_oversight_analyzer.log")
21
  logger = logging.getLogger(__name__)
22
 
23
  # Persistent directory
 
34
  os.makedirs(directory, exist_ok=True)
35
 
36
  os.environ["HF_HOME"] = model_cache_dir
 
37
  os.environ["VLLM_CACHE_DIR"] = vllm_cache_dir
38
  os.environ["TOKENIZERS_PARALLELISM"] = "false"
39
  os.environ["CUDA_LAUNCH_BLOCKING"] = "1"
40
+ # Remove TRANSFORMERS_CACHE to suppress warning
41
+ if "TRANSFORMERS_CACHE" in os.environ:
42
+ del os.environ["TRANSFORMERS_CACHE"]
43
 
44
  current_dir = os.path.dirname(os.path.abspath(__file__))
45
  src_path = os.path.abspath(os.path.join(current_dir, "src"))
 
197
  tool_files_dict={"new_tool": target_tool_path},
198
  force_finish=True,
199
  enable_checker=True,
200
+ step_rag_num=1,
201
  seed=100,
202
  additional_default_tools=[],
 
203
  )
204
  agent.init_model()
205
  log_system_usage("After Load")
 
231
  message=prompt,
232
  history=[],
233
  temperature=0.2,
234
+ max_new_tokens=256,
235
+ max_token=1024,
236
  call_agent=False,
237
  conversation=[],
238
  ):