Update app.py
Browse files
app.py
CHANGED
@@ -3,9 +3,9 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline
|
|
3 |
|
4 |
app = FastAPI()
|
5 |
|
6 |
-
# π
|
7 |
MODEL_NAME = "rinna/japanese-gpt-neox-3.6b"
|
8 |
-
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
|
9 |
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
|
10 |
generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
11 |
|
@@ -17,3 +17,4 @@ def root():
|
|
17 |
def generate(prompt: str, max_length: int = 100):
|
18 |
result = generator(prompt, max_length=max_length, do_sample=True)[0]['generated_text']
|
19 |
return {"response": result}
|
|
|
|
3 |
|
4 |
app = FastAPI()
|
5 |
|
6 |
+
# π θ»½ιγͺζ₯ζ¬θͺγ’γγ«γιΈζοΌ7B β 3.6BοΌ
|
7 |
MODEL_NAME = "rinna/japanese-gpt-neox-3.6b"
|
8 |
+
tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME, use_fast=False)
|
9 |
model = AutoModelForCausalLM.from_pretrained(MODEL_NAME)
|
10 |
generator = pipeline("text-generation", model=model, tokenizer=tokenizer)
|
11 |
|
|
|
17 |
def generate(prompt: str, max_length: int = 100):
|
18 |
result = generator(prompt, max_length=max_length, do_sample=True)[0]['generated_text']
|
19 |
return {"response": result}
|
20 |
+
|