Update app.py
Browse files
app.py
CHANGED
@@ -24,6 +24,7 @@ base_model = AutoPeftModelForCausalLM.from_pretrained(
|
|
24 |
)
|
25 |
"""
|
26 |
model = AutoPeftModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
|
|
|
27 |
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
28 |
tokenizer.pad_token_id = tokenizer.eos_token_id
|
29 |
|
@@ -54,7 +55,8 @@ train_data = train_data["train"]
|
|
54 |
def generate_answer(question):
|
55 |
#inputs = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
|
56 |
inputs = tokenizer(question, return_tensors="pt")
|
57 |
-
|
|
|
58 |
answer = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
59 |
return answer
|
60 |
|
@@ -63,8 +65,8 @@ iface = gr.Interface(
|
|
63 |
inputs="text",
|
64 |
outputs="text",
|
65 |
title="Straight Outta Logs",
|
66 |
-
examples = [prepare_sample_text(test_data[0], tokenizer, start=0, end=3), prepare_sample_text(test_data[4], tokenizer, start=0, end=5), prepare_sample_text(test_data[50], tokenizer, start=0, end=1)]
|
67 |
-
description="Use the examples or copy own
|
68 |
)
|
69 |
|
70 |
iface.launch(share=True) # Deploy the interface
|
|
|
24 |
)
|
25 |
"""
|
26 |
model = AutoPeftModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
|
27 |
+
model.to("cuda")
|
28 |
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
29 |
tokenizer.pad_token_id = tokenizer.eos_token_id
|
30 |
|
|
|
55 |
def generate_answer(question):
|
56 |
#inputs = tokenizer.apply_chat_template(messages, tokenize=True, add_generation_prompt=True, return_tensors="pt")
|
57 |
inputs = tokenizer(question, return_tensors="pt")
|
58 |
+
inputs.to("cuda")
|
59 |
+
outputs = model.generate(**inputs, max_length=2048, num_return_sequences=1, do_sample=True)
|
60 |
answer = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
61 |
return answer
|
62 |
|
|
|
65 |
inputs="text",
|
66 |
outputs="text",
|
67 |
title="Straight Outta Logs",
|
68 |
+
examples = [prepare_sample_text(test_data[0], tokenizer, start=0, end=3), prepare_sample_text(test_data[4], tokenizer, start=0, end=5), prepare_sample_text(test_data[50], tokenizer, start=0, end=1)],
|
69 |
+
description="Use the examples or copy own sepsis case example",
|
70 |
)
|
71 |
|
72 |
iface.launch(share=True) # Deploy the interface
|