File size: 1,612 Bytes
576fb99
 
 
ecc70cc
0dc21cf
ecc70cc
 
 
 
 
23fe41c
576fb99
 
 
57fa22a
 
725414f
 
 
 
57fa22a
725414f
 
57fa22a
725414f
57fa22a
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
576fb99
57fa22a
 
 
 
725414f
 
 
 
57fa22a
725414f
 
57fa22a
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
from transformers import AutoTokenizer, AutoModelForSeq2SeqLM



checkpoint = "https://huggingface.co/Shivam29rathore/t5_10k_base"

#load model from drive
#with open('Shivam29rathore/t5_10k_base', "rb") as f:
  #t5 =  pickle.load(f)

#tokenizer = AutoTokenizer.from_pretrained(checkpoint)
model = AutoModelForSeq2SeqLM.from_pretrained(checkpoint)


def summarize(word):
    import os
    data_path = "/tmp/"
    if not os.path.exists(data_path):
        os.makedirs(data_path)
    input_ = "/tmp/input.txt"

    with open(input_, "w") as file:
        file.write(word)
    # read the written txt into a variable
    with open(input_ , 'r') as f:
        text_ = f.read()

    def clean_data(texts):
        import re
        words = list()
        for text in texts.split():
            text = re.sub(r'\n','',text)
            text = re.sub(r'\s$','',text)
            words.append(text)

        return "summarize " + " ".join(words)
    text = clean_data(text_)

    final_summary = []
    for x in range(0,len(text)-1256,1256):
        text_to_summarize= text[x:x+1256]
        final_summary.append(model.predict(text_to_summarize))

    final_list = list(itertools.chain.from_iterable(final_summary))
    final_list = ''.join(final_list)
    return final_list
    
    
import gradio as gr

iface = gr.Interface(fn= summarize, 
                     inputs =gr.inputs.Textbox(lines=15,placeholder="Enter your text !!"),
                     outputs="text",title="Document Summarizer",description ="An AI that makes your life easier by helping you summarise long texts.")
iface.launch(auth=("docai","ailabs"))