TalkGPT

This model is a fine-tuned version of BlenderBot-400M (distilled) based on a custom conversational dataset. It is designed to generate conversational responses in English.

License

Apache 2.0

Datasets

The model is fine-tuned on a custom dataset consisting of conversational dialogues.

Language

English

Metrics

  • BLEU: 0.1687 (calculated on the validation set)
  • ROUGE-1: 0.4078
  • ROUGE-2: 0.1912
  • ROUGE-L: 0.3418
  • ROUGE-Lsum: 0.3401
  • Training Loss: 0.2460 (final training loss after fine-tuning)

Base Model

The model is based on the BlenderBot-400M-distill architecture by Facebook AI.

Pipeline Tag

text-generation

Library Name

transformers

Tags

BlenderBot, Conversational, Fine-tuned, Text Generation

Eval Results

The model achieved the following results on the validation set:

  • BLEU: 0.1687
  • ROUGE-1: 0.4078
  • ROUGE-2: 0.1912
  • ROUGE-L: 0.3418
  • ROUGE-Lsum: 0.3401
  • Training Loss: 0.2460 after 3 epochs of fine-tuning.
Downloads last month
110
Safetensors
Model size
365M params
Tensor type
F32
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for 12sciencejnv/TalkGPT

Finetuned
(16)
this model

Evaluation results