TKgumi commited on
Commit
1ad1638
Β·
verified Β·
1 Parent(s): c7798b8

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -2
app.py CHANGED
@@ -3,9 +3,9 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
3
 
4
  app = FastAPI()
5
 
6
- # πŸš€ γ•γ‚‰γ«θ»½γ„γƒ’γƒ‡γƒ«γ‚’ιΈζŠžοΌˆ7B β†’ 3.6BοΌ‰
7
  MODEL_NAME = "rinna/japanese-gpt-neox-3.6b"
8
- tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
9
  model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
10
  generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
11
 
@@ -17,3 +17,4 @@ def root():
17
  def generate(prompt: str, max_length: int = 100):
18
  result = generator(prompt, max_length=max_length, do_sample=True)[0]['generated_text']
19
  return {"response": result}
 
 
3
 
4
  app = FastAPI()
5
 
6
+ # πŸš€ 軽量γͺζ—₯本θͺžγƒ’γƒ‡γƒ«γ‚’ιΈζŠžοΌˆ7B β†’ 3.6BοΌ‰
7
  MODEL_NAME = "rinna/japanese-gpt-neox-3.6b"
8
+ tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=False)
9
  model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
10
  generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
11
 
 
17
  def generate(prompt: str, max_length: int = 100):
18
  result = generator(prompt, max_length=max_length, do_sample=True)[0]['generated_text']
19
  return {"response": result}
20
+