arya-ai-model commited on
Commit
a983f3b
·
1 Parent(s): 0512849

First commit

Browse files
Files changed (2) hide show
  1. Dockerfile +4 -0
  2. app.py +6 -4
Dockerfile CHANGED
@@ -6,4 +6,8 @@ RUN pip install --no-cache-dir -r requirements.txt
6
 
7
  COPY . .
8
 
 
 
 
 
9
  CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
 
6
 
7
  COPY . .
8
 
9
+ mkdir -p /tmp/huggingface
10
+ chmod -R 777 /tmp/huggingface
11
+
12
+
13
  CMD ["uvicorn", "app:app", "--host", "0.0.0.0", "--port", "7860"]
app.py CHANGED
@@ -1,13 +1,15 @@
1
  import os
 
 
 
 
 
 
2
  import torch
3
  from fastapi import FastAPI, HTTPException
4
  from pydantic import BaseModel
5
  from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
6
 
7
- # Set custom cache directory
8
- os.environ["HF_HOME"] = "/tmp/huggingface"
9
- os.environ["TRANSFORMERS_CACHE"] = "/tmp/huggingface"
10
-
11
  # Model setup
12
  MODEL_NAME = "deepseek-ai/deepseek-llm-7b-base"
13
  DEVICE = "cuda" if torch.cuda.is_available() else "cpu"
 
1
  import os
2
+
3
+ # Set a writable cache directory
4
+ os.environ["HF_HOME"] = "/tmp/huggingface"
5
+ os.environ["TRANSFORMERS_CACHE"] = "/tmp/huggingface"
6
+
7
+ # Now import the required libraries
8
  import torch
9
  from fastapi import FastAPI, HTTPException
10
  from pydantic import BaseModel
11
  from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig
12
 
 
 
 
 
13
  # Model setup
14
  MODEL_NAME = "deepseek-ai/deepseek-llm-7b-base"
15
  DEVICE = "cuda" if torch.cuda.is_available() else "cpu"