Uhhy commited on
Commit
5726c3f
·
verified ·
1 Parent(s): 44fd78a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -6
app.py CHANGED
@@ -2,10 +2,10 @@ from fastapi import FastAPI, HTTPException
2
  from pydantic import BaseModel
3
  from llama_cpp import Llama
4
  from concurrent.futures import ThreadPoolExecutor, as_completed
5
- from tqdm import tqdm
6
  import uvicorn
7
  from dotenv import load_dotenv
8
  import re
 
9
  import spaces
10
 
11
  load_dotenv()
@@ -59,7 +59,7 @@ class ModelManager:
59
  try:
60
  model = future.result()
61
  models.append(model)
62
- except Exception as e:
63
  pass
64
 
65
  self.models = models
@@ -67,7 +67,6 @@ class ModelManager:
67
  return self.models
68
 
69
  model_manager = ModelManager()
70
-
71
  global_data['models'] = model_manager.load_all_models()
72
 
73
  class ChatRequest(BaseModel):
@@ -130,11 +129,11 @@ async def generate_chat(request: ChatRequest):
130
 
131
  with ThreadPoolExecutor() as executor:
132
  futures = [executor.submit(generate_chat_response, request, model_data) for model_data in global_data['models']]
133
- for future in tqdm(as_completed(futures), total=num_models):
134
  try:
135
  response = future.result()
136
  responses.append(response)
137
- except Exception as exc:
138
  pass
139
 
140
  if not responses:
@@ -148,4 +147,4 @@ async def generate_chat(request: ChatRequest):
148
  }
149
 
150
  if __name__ == "__main__":
151
- uvicorn.run(app, host="0.0.0.0", port=8000)
 
2
  from pydantic import BaseModel
3
  from llama_cpp import Llama
4
  from concurrent.futures import ThreadPoolExecutor, as_completed
 
5
  import uvicorn
6
  from dotenv import load_dotenv
7
  import re
8
+ import huggingface_hub
9
  import spaces
10
 
11
  load_dotenv()
 
59
  try:
60
  model = future.result()
61
  models.append(model)
62
+ except Exception:
63
  pass
64
 
65
  self.models = models
 
67
  return self.models
68
 
69
  model_manager = ModelManager()
 
70
  global_data['models'] = model_manager.load_all_models()
71
 
72
  class ChatRequest(BaseModel):
 
129
 
130
  with ThreadPoolExecutor() as executor:
131
  futures = [executor.submit(generate_chat_response, request, model_data) for model_data in global_data['models']]
132
+ for future in as_completed(futures):
133
  try:
134
  response = future.result()
135
  responses.append(response)
136
+ except Exception:
137
  pass
138
 
139
  if not responses:
 
147
  }
148
 
149
  if __name__ == "__main__":
150
+ uvicorn.run(app, host="0.0.0.0", port=7860)