summarisation_model / README.md
Saravanankumaran's picture
Update README.md
f4ac15a verified
metadata
library_name: transformers
tags:
  - summarization
  - transformers
  - t5
  - fine-tuning
  - custom-dataset
  - text-generation

Model Card for Model ID

Model Details

Model Description

This is a fine-tuned T5 model for text summarization using the SAMSum dataset. The model has been trained using 🤗 Transformers and Hugging Face Trainer with mixed precision (fp16) to optimize memory efficiency.

  • Developed by: Saravanan K
  • Finetuned from model [optional]: t5-base

Model Sources [optional]

Use Cases

Direct Use

This model can be used for text summarization tasks, particularly for summarizing dialogues and conversations.

Downstream Use

The model can be fine-tuned further on other summarization datasets or used in larger NLP applications requiring summarization capabilities.

Out Of Scope Use

The model may not perform well on non-dialogue-based text or non-English languages.

Bias, Risks, and Limitations

Biases: Since it is trained on the SAMSum dataset, it may have biases related to conversational English data. Limitations: Performance may degrade on texts that are significantly different from the training dataset.

How to Get Started with the Model

Use the code below to get started with the model.

from transformers import AutoModelForSeq2SeqLM, AutoTokenizer

Define model name (same as uploaded one)

model_name = "Saravanankumaran/summarisation_model"

Load model and tokenizer from Hugging Face Hub

model = AutoModelForSeq2SeqLM.from_pretrained(model_name) tokenizer = AutoTokenizer.from_pretrained(model_name)

print("Model loaded successfully! ✅")

Below is the example text to use the model

text = """ Laxmi Kant: what work you planning to give Tom? Juli: i was hoping to send him on a business trip first. Laxmi Kant: cool. is there any suitable work for him? Juli: he did excellent in last quarter. i will assign new project, once he is back. """ inputs = tokenizer(text, return_tensors="pt")

output = model.generate(**inputs) summary = tokenizer.decode(output[0], skip_special_tokens=True)

print("Generated Output:", summary)