update demo link
Browse files
README.md
CHANGED
@@ -280,15 +280,14 @@ model-index:
|
|
280 |
|
281 |
# Longformer Encoder-Decoder (LED) fine-tuned on Booksum
|
282 |
|
283 |
-
demo:
|
284 |
|
285 |
-
[![colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/gist/pszemraj/
|
286 |
|
287 |
- A fine-tuned version of [allenai/led-large-16384](https://huggingface.co/allenai/led-large-16384) on the BookSum dataset.
|
288 |
-
- Goal: a model that can generalize well and is useful in summarizing long text in academic and daily usage.
|
289 |
- works well on lots of text and can handle 16384 tokens/batch (_if you have the GPU memory to handle that_)
|
290 |
|
291 |
-
> Note: the API is set to generate a max of 64 tokens for runtime reasons, so the summaries may be truncated (depending on length of input text). For best results use python as below.
|
292 |
|
293 |
---
|
294 |
|
@@ -366,7 +365,7 @@ The following hyperparameters were used during training:
|
|
366 |
|
367 |
#### In-between Epochs
|
368 |
|
369 |
-
Unfortunately, don't have all records on-hand for middle epochs
|
370 |
|
371 |
- learning_rate: 4e-05
|
372 |
- train_batch_size: 2
|
|
|
280 |
|
281 |
# Longformer Encoder-Decoder (LED) fine-tuned on Booksum
|
282 |
|
|
|
283 |
|
284 |
+
[![colab](https://colab.research.google.com/assets/colab-badge.svg)](https://colab.research.google.com/gist/pszemraj/3eba944ddc9fc9a4a1bfb21e83b57620/summarization-token-batching.ipynb)
|
285 |
|
286 |
- A fine-tuned version of [allenai/led-large-16384](https://huggingface.co/allenai/led-large-16384) on the BookSum dataset.
|
287 |
+
- Goal: a model that can generalize well and is useful in summarizing long text in academic and daily usage. See the demo linked above!
|
288 |
- works well on lots of text and can handle 16384 tokens/batch (_if you have the GPU memory to handle that_)
|
289 |
|
290 |
+
> Note: the API is set to generate a max of 64 tokens for runtime reasons, so the summaries may be truncated (depending on the length of input text). For best results use python as below.
|
291 |
|
292 |
---
|
293 |
|
|
|
365 |
|
366 |
#### In-between Epochs
|
367 |
|
368 |
+
Unfortunately, don't have all records on-hand for middle epochs; the following should be representative:
|
369 |
|
370 |
- learning_rate: 4e-05
|
371 |
- train_batch_size: 2
|