BART Financial Summarization Model

Model Name: kritsadaK/bart-financial-summarization
Base Model: facebook/bart-large-cnn
Task: Financial Text Summarization
Dataset: kritsadaK/EDGAR-CORPUS-Financial-Summarization

Techniques:

  • Fine-tuned using the Hugging Face Trainer API
  • Tokenized with AutoTokenizer (max length 1024 for input, 256 for summary)
  • Optimized with AdamW, learning rate 2e-5, batch size 2, fp16 enabled
  • Evaluated using ROUGE scores

Evaluation Results:

  • Loss: 1.18
  • Runtime: 18.9 seconds
  • Samples per second: 56.1
  • Steps per second: 28.1
  • Epochs: 3

Usage Example (Python):

from transformers import pipeline

summarizer = pipeline("summarization", model="kritsadaK/bart-financial-summarization")
text = "Your financial document text here..."
summary = summarizer(text, max_length=256, min_length=50, do_sample=False)
print(summary)

The Financial Statements Summary 10K Dataset was developed as part of the CSX4210: Natural Language Processing project at Assumption University.

Downloads last month
23
Safetensors
Model size
406M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Model tree for kritsadaK/bart-financial-summarization

Finetuned
(334)
this model

Dataset used to train kritsadaK/bart-financial-summarization