Spaces:
Build error
Build error
Update app.py
Browse files
app.py
CHANGED
@@ -2,10 +2,10 @@ from fastapi import FastAPI, HTTPException
|
|
2 |
from pydantic import BaseModel
|
3 |
from llama_cpp import Llama
|
4 |
from concurrent.futures import ThreadPoolExecutor, as_completed
|
5 |
-
from tqdm import tqdm
|
6 |
import uvicorn
|
7 |
from dotenv import load_dotenv
|
8 |
import re
|
|
|
9 |
import spaces
|
10 |
|
11 |
load_dotenv()
|
@@ -59,7 +59,7 @@ class ModelManager:
|
|
59 |
try:
|
60 |
model = future.result()
|
61 |
models.append(model)
|
62 |
-
except Exception
|
63 |
pass
|
64 |
|
65 |
self.models = models
|
@@ -67,7 +67,6 @@ class ModelManager:
|
|
67 |
return self.models
|
68 |
|
69 |
model_manager = ModelManager()
|
70 |
-
|
71 |
global_data['models'] = model_manager.load_all_models()
|
72 |
|
73 |
class ChatRequest(BaseModel):
|
@@ -130,11 +129,11 @@ async def generate_chat(request: ChatRequest):
|
|
130 |
|
131 |
with ThreadPoolExecutor() as executor:
|
132 |
futures = [executor.submit(generate_chat_response, request, model_data) for model_data in global_data['models']]
|
133 |
-
for future in
|
134 |
try:
|
135 |
response = future.result()
|
136 |
responses.append(response)
|
137 |
-
except Exception
|
138 |
pass
|
139 |
|
140 |
if not responses:
|
@@ -148,4 +147,4 @@ async def generate_chat(request: ChatRequest):
|
|
148 |
}
|
149 |
|
150 |
if __name__ == "__main__":
|
151 |
-
uvicorn.run(app, host="0.0.0.0", port=
|
|
|
2 |
from pydantic import BaseModel
|
3 |
from llama_cpp import Llama
|
4 |
from concurrent.futures import ThreadPoolExecutor, as_completed
|
|
|
5 |
import uvicorn
|
6 |
from dotenv import load_dotenv
|
7 |
import re
|
8 |
+
import huggingface_hub
|
9 |
import spaces
|
10 |
|
11 |
load_dotenv()
|
|
|
59 |
try:
|
60 |
model = future.result()
|
61 |
models.append(model)
|
62 |
+
except Exception:
|
63 |
pass
|
64 |
|
65 |
self.models = models
|
|
|
67 |
return self.models
|
68 |
|
69 |
model_manager = ModelManager()
|
|
|
70 |
global_data['models'] = model_manager.load_all_models()
|
71 |
|
72 |
class ChatRequest(BaseModel):
|
|
|
129 |
|
130 |
with ThreadPoolExecutor() as executor:
|
131 |
futures = [executor.submit(generate_chat_response, request, model_data) for model_data in global_data['models']]
|
132 |
+
for future in as_completed(futures):
|
133 |
try:
|
134 |
response = future.result()
|
135 |
responses.append(response)
|
136 |
+
except Exception:
|
137 |
pass
|
138 |
|
139 |
if not responses:
|
|
|
147 |
}
|
148 |
|
149 |
if __name__ == "__main__":
|
150 |
+
uvicorn.run(app, host="0.0.0.0", port=7860)
|