File size: 768 Bytes
568a3e4
 
142690b
568a3e4
 
 
 
 
 
142690b
568a3e4
 
 
 
142690b
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
from transformers import AutoModelForMaskedLM , AutoTokenizer
import torch
model_path="bert-base-multilingual-uncased"
tokenizer = AutoTokenizer.from_pretrained(model_path)
# load Prompting class
from prompt import Prompting
prompting= Prompting(model=model_path)
prompt= ". Because it was "+ prompting.tokenizer.mask_token +"."


def predict(text):
  THRESHOLD = prompting.compute_tokens_prob(prompt, token_list1=["good"], token_list2= ["bad"])[0].item()
  res=prompting.compute_tokens_prob(text+prompt, token_list1=["good"], token_list2= ["bad"])
  if res[0] > THRESHOLD:
    return {"POSITIVE":res[0].item()-THRESHOLD}
  return {"NEGATIVE":THRESHOLD-res[1].item()}



import gradio as gr
iface = gr.Interface(fn=predict, inputs=["text"], outputs=["label"]).launch()