metadata
language:
- en
thumbnail: null
tags:
- pytorch
- google/pegasus-reddit_tifu
- summarization
- samsum
license: null
datasets:
- samsum
metrics:
- rouge
Samsum Pegasus (Reddit/TIFU) for conversational summaries
Model description
Pegasus (Reddit/TIFU) for conversational summaries trained on the samsum dataset!
Training data
The data is the samsum dataset for conversional summaries.
The initial weigths were from the google/pegasus-reddit_tifu. The hypothesis being that it would help the convergence on the samsum dataset to have weights trained on a larger summarization dataset first like the Reddit TIFU using casual language.
Training procedure
Used the example/seq2seq/run_summarization.py script from the transformers source 4.5.0dev0.
n_epochs: 3,
batch_size: 4,
max_source_length: 512,
max_target_length: 128
Eval results
eval_gen_len: 35.89,
eval_loss: 1.3807392120361328,
eval_rouge1: 47.3372,
eval_rouge2: 24.4728,
eval_rougeL: 37.9078,
eval_rougeLsum: 43.5744,
eval_samples_per_second: 2.814