Pratham Bhat commited on
Commit
83a36ef
·
1 Parent(s): 853e734

Changed auto tokenizer to granite tokenizer

Browse files
Files changed (1) hide show
  1. main.py +2 -2
main.py CHANGED
@@ -11,7 +11,7 @@
11
 
12
  from fastapi import FastAPI
13
  from pydantic import BaseModel
14
- from transformers import AutoModelForCausalLM, AutoTokenizer
15
  import uvicorn
16
  import torch
17
  # torch.mps.empty_cache()
@@ -62,7 +62,7 @@ def generate(item: Item):
62
  device = "cuda" if torch.cuda.is_available() else "cpu"
63
 
64
  model_path = "ibm-granite/granite-34b-code-instruct-8k"
65
- tokenizer = AutoTokenizer.from_pretrained(model_path)
66
  # drop device_map if running on CPU
67
  model = AutoModelForCausalLM.from_pretrained(model_path, device_map=device)
68
  model.eval()
 
11
 
12
  from fastapi import FastAPI
13
  from pydantic import BaseModel
14
+ from transformers import AutoModelForCausalLM, AutoTokenizer, GraniteTokenizer
15
  import uvicorn
16
  import torch
17
  # torch.mps.empty_cache()
 
62
  device = "cuda" if torch.cuda.is_available() else "cpu"
63
 
64
  model_path = "ibm-granite/granite-34b-code-instruct-8k"
65
+ tokenizer = GraniteTokenizer.from_pretrained(model_path)
66
  # drop device_map if running on CPU
67
  model = AutoModelForCausalLM.from_pretrained(model_path, device_map=device)
68
  model.eval()