Spaces:
Sleeping
Sleeping
File size: 721 Bytes
1856062 e5413bc 3cc14e8 576bbe0 3cc14e8 576bbe0 3cc14e8 576bbe0 3cc14e8 576bbe0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 |
from fastapi import FastAPI
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
app = FastAPI()
@app.get("/")
def greet_json():
return {"Hello": "World!"}
@app.get("/hello/{msg}")
def say_hello(msg: str):
print("model")
tokenizer = AutoTokenizer.from_pretrained("google/gemma-2b-it")
model = AutoModelForCausalLM.from_pretrained(
"google/gemma-2b-it",
device_map="auto",
torch_dtype=torch.bfloat16
)
print("token & msg")
input_ids = tokenizer(msg, return_tensors="pt").to("cpu")
print("output")
outputs = model.generate(**input_ids, max_length=500)
print("complete")
return {"message": tokenizer.decode(outputs[0])} |