arya-ai-model commited on
Commit
d9ded01
·
1 Parent(s): 22b2a0c

updated model.py

Browse files
Files changed (1) hide show
  1. model.py +14 -5
model.py CHANGED
@@ -1,16 +1,25 @@
 
1
  from transformers import AutoModelForCausalLM, AutoTokenizer
2
  import torch
3
 
4
- # Load the pre-trained model
5
- MODEL_NAME = "starcoder"
6
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
7
- model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, device_map="auto")
 
 
 
 
 
 
 
8
 
9
  def generate_code(prompt: str, max_tokens: int = 256):
10
  """Generates code based on the input prompt."""
11
  if not prompt.strip():
12
  return "Error: Empty prompt provided."
13
 
14
- inputs = tokenizer(prompt, return_tensors="pt").to("cuda" if torch.cuda.is_available() else "cpu")
 
15
  output = model.generate(**inputs, max_new_tokens=max_tokens)
16
  return tokenizer.decode(output[0], skip_special_tokens=True)
 
1
+ import os
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
  import torch
4
 
5
+ # Correct model name
6
+ MODEL_NAME = "bigcode/starcoder"
7
+
8
+ # Ensure the token is provided
9
+ HF_TOKEN = os.getenv("HUGGINGFACE_TOKEN")
10
+ if not HF_TOKEN:
11
+ raise ValueError("Missing Hugging Face token. Set HUGGINGFACE_TOKEN as an environment variable.")
12
+
13
+ # Load tokenizer and model with authentication
14
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, token=HF_TOKEN)
15
+ model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, token=HF_TOKEN, device_map="auto")
16
 
17
  def generate_code(prompt: str, max_tokens: int = 256):
18
  """Generates code based on the input prompt."""
19
  if not prompt.strip():
20
  return "Error: Empty prompt provided."
21
 
22
+ device = "cuda" if torch.cuda.is_available() else "cpu"
23
+ inputs = tokenizer(prompt, return_tensors="pt").to(device)
24
  output = model.generate(**inputs, max_new_tokens=max_tokens)
25
  return tokenizer.decode(output[0], skip_special_tokens=True)