Spaces:
Build error
Build error
Pratham Bhat
commited on
Commit
·
83a36ef
1
Parent(s):
853e734
Changed auto tokenizer to granite tokenizer
Browse files
main.py
CHANGED
@@ -11,7 +11,7 @@
|
|
11 |
|
12 |
from fastapi import FastAPI
|
13 |
from pydantic import BaseModel
|
14 |
-
from transformers import AutoModelForCausalLM, AutoTokenizer
|
15 |
import uvicorn
|
16 |
import torch
|
17 |
# torch.mps.empty_cache()
|
@@ -62,7 +62,7 @@ def generate(item: Item):
|
|
62 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
63 |
|
64 |
model_path = "ibm-granite/granite-34b-code-instruct-8k"
|
65 |
-
tokenizer =
|
66 |
# drop device_map if running on CPU
|
67 |
model = AutoModelForCausalLM.from_pretrained(model_path, device_map=device)
|
68 |
model.eval()
|
|
|
11 |
|
12 |
from fastapi import FastAPI
|
13 |
from pydantic import BaseModel
|
14 |
+
from transformers import AutoModelForCausalLM, AutoTokenizer, GraniteTokenizer
|
15 |
import uvicorn
|
16 |
import torch
|
17 |
# torch.mps.empty_cache()
|
|
|
62 |
device = "cuda" if torch.cuda.is_available() else "cpu"
|
63 |
|
64 |
model_path = "ibm-granite/granite-34b-code-instruct-8k"
|
65 |
+
tokenizer = GraniteTokenizer.from_pretrained(model_path)
|
66 |
# drop device_map if running on CPU
|
67 |
model = AutoModelForCausalLM.from_pretrained(model_path, device_map=device)
|
68 |
model.eval()
|