Spaces:
Runtime error
Runtime error
Commit
·
a983f3b
1
Parent(s):
0512849
First commit
Browse files- Dockerfile +4 -0
- app.py +6 -4
Dockerfile
CHANGED
@@ -6,4 +6,8 @@ RUN pip install --no-cache-dir -r requirements.txt
|
|
6 |
|
7 |
COPY . .
|
8 |
|
|
|
|
|
|
|
|
|
9 |
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
|
|
|
6 |
|
7 |
COPY . .
|
8 |
|
9 |
+
mkdir -p /tmp/huggingface
|
10 |
+
chmod -R 777 /tmp/huggingface
|
11 |
+
|
12 |
+
|
13 |
CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
|
app.py
CHANGED
@@ -1,13 +1,15 @@
|
|
1 |
import os
|
|
|
|
|
|
|
|
|
|
|
|
|
2 |
import torch
|
3 |
from fastapi import FastAPI, HTTPException
|
4 |
from pydantic import BaseModel
|
5 |
from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
|
6 |
|
7 |
-
# Set custom cache directory
|
8 |
-
os.environ["HF_HOME"] = "/tmp/huggingface"
|
9 |
-
os.environ["TRANSFORMERS_CACHE"] = "/tmp/huggingface"
|
10 |
-
|
11 |
# Model setup
|
12 |
MODEL_NAME = "deepseek-ai/deepseek-llm-7b-base"
|
13 |
DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
|
|
|
1 |
import os
|
2 |
+
|
3 |
+
# Set a writable cache directory
|
4 |
+
os.environ["HF_HOME"] = "/tmp/huggingface"
|
5 |
+
os.environ["TRANSFORMERS_CACHE"] = "/tmp/huggingface"
|
6 |
+
|
7 |
+
# Now import the required libraries
|
8 |
import torch
|
9 |
from fastapi import FastAPI, HTTPException
|
10 |
from pydantic import BaseModel
|
11 |
from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
|
12 |
|
|
|
|
|
|
|
|
|
13 |
# Model setup
|
14 |
MODEL_NAME = "deepseek-ai/deepseek-llm-7b-base"
|
15 |
DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
|