Update app.py
Browse files
app.py
CHANGED
@@ -20,12 +20,11 @@ logging.basicConfig(
|
|
20 |
|
21 |
def proc_submission(
|
22 |
input_text: str,
|
23 |
-
model_type: str,
|
24 |
-
summary_type: str,
|
25 |
num_beams,
|
26 |
token_batch_length,
|
27 |
length_penalty,
|
28 |
-
#repetition_penalty,
|
29 |
#no_repeat_ngram_size: int = 3,
|
30 |
max_input_length: int = 768,
|
31 |
):
|
@@ -71,7 +70,7 @@ def proc_submission(
|
|
71 |
tr_in = input_text
|
72 |
msg = None
|
73 |
|
74 |
-
if
|
75 |
_summaries = summarize_via_tokenbatches(
|
76 |
tr_in,
|
77 |
model_led_tldr if (summary_type == "TLDR") else model_led_det,
|
@@ -79,7 +78,7 @@ def proc_submission(
|
|
79 |
batch_length=token_batch_length,
|
80 |
**settings,
|
81 |
)
|
82 |
-
elif
|
83 |
_summaries = summarize_via_tokenbatches(
|
84 |
tr_in,
|
85 |
model_tldr if (summary_type == "TLDR") else model_det,
|
@@ -87,13 +86,13 @@ def proc_submission(
|
|
87 |
batch_length=token_batch_length,
|
88 |
**settings,
|
89 |
)
|
90 |
-
_summaries = summarize_via_tokenbatches(
|
91 |
-
tr_in,
|
92 |
-
model_tldr if (summary_type == "TLDR") else model_det,
|
93 |
-
tokenizer_tldr if (summary_type == "TLDR") else tokenizer_det,
|
94 |
-
batch_length=token_batch_length,
|
95 |
-
|
96 |
-
)
|
97 |
sum_text = [f"Section {i}: " + s["summary"][0] for i, s in enumerate(_summaries)]
|
98 |
compression_rate = [
|
99 |
f" - Section {i}: {round(s['compression_rate'],3)}"
|
|
|
20 |
|
21 |
def proc_submission(
|
22 |
input_text: str,
|
23 |
+
model_type: str = 'LongT5',
|
24 |
+
summary_type: str = 'TLDR',
|
25 |
num_beams,
|
26 |
token_batch_length,
|
27 |
length_penalty,
|
|
|
28 |
#no_repeat_ngram_size: int = 3,
|
29 |
max_input_length: int = 768,
|
30 |
):
|
|
|
70 |
tr_in = input_text
|
71 |
msg = None
|
72 |
|
73 |
+
if model_type == "LED":
|
74 |
_summaries = summarize_via_tokenbatches(
|
75 |
tr_in,
|
76 |
model_led_tldr if (summary_type == "TLDR") else model_led_det,
|
|
|
78 |
batch_length=token_batch_length,
|
79 |
**settings,
|
80 |
)
|
81 |
+
elif model_type == "LongT5":
|
82 |
_summaries = summarize_via_tokenbatches(
|
83 |
tr_in,
|
84 |
model_tldr if (summary_type == "TLDR") else model_det,
|
|
|
86 |
batch_length=token_batch_length,
|
87 |
**settings,
|
88 |
)
|
89 |
+
#_summaries = summarize_via_tokenbatches(
|
90 |
+
#tr_in,
|
91 |
+
#model_tldr if (summary_type == "TLDR") else model_det,
|
92 |
+
#tokenizer_tldr if (summary_type == "TLDR") else tokenizer_det,
|
93 |
+
#batch_length=token_batch_length,
|
94 |
+
#**settings,
|
95 |
+
#)
|
96 |
sum_text = [f"Section {i}: " + s["summary"][0] for i, s in enumerate(_summaries)]
|
97 |
compression_rate = [
|
98 |
f" - Section {i}: {round(s['compression_rate'],3)}"
|