Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -7,21 +7,26 @@ from transformers import (
|
|
7 |
from threading import Thread
|
8 |
|
9 |
# Configuration
|
10 |
-
MODEL_NAME = "deepseek-ai/DeepSeek-R1"
|
11 |
DEFAULT_MAX_NEW_TOKENS = 512
|
12 |
|
13 |
-
# Load model and tokenizer
|
14 |
try:
|
15 |
-
tokenizer = AutoTokenizer.from_pretrained(
|
|
|
|
|
|
|
16 |
model = AutoModelForCausalLM.from_pretrained(
|
17 |
MODEL_NAME,
|
18 |
device_map="auto",
|
19 |
torch_dtype="auto",
|
20 |
-
|
|
|
21 |
)
|
22 |
except Exception as e:
|
23 |
raise gr.Error(f"Error loading model: {str(e)}")
|
24 |
|
|
|
25 |
def generate_text(prompt, max_new_tokens=DEFAULT_MAX_NEW_TOKENS, temperature=0.7, top_p=0.9):
|
26 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
27 |
|
|
|
7 |
from threading import Thread
|
8 |
|
9 |
# Configuration
|
10 |
+
MODEL_NAME = "deepseek-ai/DeepSeek-R1"
|
11 |
DEFAULT_MAX_NEW_TOKENS = 512
|
12 |
|
13 |
+
# Load model and tokenizer WITH TRUSTED CODE
|
14 |
try:
|
15 |
+
tokenizer = AutoTokenizer.from_pretrained(
|
16 |
+
MODEL_NAME,
|
17 |
+
trust_remote_code=True # <-- ADDED HERE
|
18 |
+
)
|
19 |
model = AutoModelForCausalLM.from_pretrained(
|
20 |
MODEL_NAME,
|
21 |
device_map="auto",
|
22 |
torch_dtype="auto",
|
23 |
+
trust_remote_code=True, # <-- ADDED HERE
|
24 |
+
# load_in_4bit=True # Uncomment for quantization
|
25 |
)
|
26 |
except Exception as e:
|
27 |
raise gr.Error(f"Error loading model: {str(e)}")
|
28 |
|
29 |
+
|
30 |
def generate_text(prompt, max_new_tokens=DEFAULT_MAX_NEW_TOKENS, temperature=0.7, top_p=0.9):
|
31 |
inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
|
32 |
|