shivrajkarewar commited on
Commit
b494df4
·
verified ·
1 Parent(s): dc1de0c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +9 -4
app.py CHANGED
@@ -7,21 +7,26 @@ from transformers import (
7
  from threading import Thread
8
 
9
  # Configuration
10
- MODEL_NAME = "deepseek-ai/DeepSeek-R1" # Verify exact model ID on Hugging Face Hub
11
  DEFAULT_MAX_NEW_TOKENS = 512
12
 
13
- # Load model and tokenizer
14
  try:
15
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
 
 
 
16
  model = AutoModelForCausalLM.from_pretrained(
17
  MODEL_NAME,
18
  device_map="auto",
19
  torch_dtype="auto",
20
- # load_in_4bit=True # Uncomment for 4-bit quantization
 
21
  )
22
  except Exception as e:
23
  raise gr.Error(f"Error loading model: {str(e)}")
24
 
 
25
  def generate_text(prompt, max_new_tokens=DEFAULT_MAX_NEW_TOKENS, temperature=0.7, top_p=0.9):
26
  inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
27
 
 
7
  from threading import Thread
8
 
9
  # Configuration
10
+ MODEL_NAME = "deepseek-ai/DeepSeek-R1"
11
  DEFAULT_MAX_NEW_TOKENS = 512
12
 
13
+ # Load model and tokenizer WITH TRUSTED CODE
14
  try:
15
+ tokenizer = AutoTokenizer.from_pretrained(
16
+ MODEL_NAME,
17
+ trust_remote_code=True # <-- ADDED HERE
18
+ )
19
  model = AutoModelForCausalLM.from_pretrained(
20
  MODEL_NAME,
21
  device_map="auto",
22
  torch_dtype="auto",
23
+ trust_remote_code=True, # <-- ADDED HERE
24
+ # load_in_4bit=True # Uncomment for quantization
25
  )
26
  except Exception as e:
27
  raise gr.Error(f"Error loading model: {str(e)}")
28
 
29
+
30
  def generate_text(prompt, max_new_tokens=DEFAULT_MAX_NEW_TOKENS, temperature=0.7, top_p=0.9):
31
  inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
32