|
import gradio as gr |
|
from huggingface_hub import from_pretrained_keras |
|
from huggingface_hub import KerasModelHubMixin |
|
import transformers |
|
from transformers import AutoTokenizer |
|
|
|
|
|
m = from_pretrained_keras('sgonzalezsilot/FakeNews-Detection-Twitter-Thesis') |
|
|
|
|
|
def bert_encode(tokenizer,data,maximum_length) : |
|
input_ids = [] |
|
attention_masks = [] |
|
|
|
|
|
for i in range(len(data)): |
|
encoded = tokenizer.encode_plus( |
|
|
|
data[i], |
|
add_special_tokens=True, |
|
max_length=maximum_length, |
|
pad_to_max_length=True, |
|
truncation = True, |
|
return_attention_mask=True, |
|
) |
|
|
|
input_ids.append(encoded['input_ids']) |
|
attention_masks.append(encoded['attention_mask']) |
|
|
|
return np.array(input_ids),np.array(attention_masks) |
|
|
|
train_encodings = tokenizer(train_texts, truncation=True, padding=True) |
|
test_encodings = tokenizer(test_texts, truncation=True, padding=True) |
|
|
|
MODEL = "digitalepidemiologylab/covid-twitter-bert-v2" |
|
tokenizer = AutoTokenizer.from_pretrained(MODEL) |
|
|
|
sentence_length = 110 |
|
train_input_ids,train_attention_masks = bert_encode(tokenizer,train_texts,sentence_length) |
|
test_input_ids,test_attention_masks = bert_encode(tokenizer,test_texts,sentence_length) |
|
|
|
def get_news(input_text): |
|
return sentiment(input_text) |
|
|
|
iface = gr.Interface(fn = get_news, |
|
inputs = "text", |
|
outputs = ['text'], |
|
title = 'Fake News', |
|
description="") |
|
|
|
iface.launch(inline = False) |