|
--- |
|
metrics: |
|
- accuracy |
|
pipeline_tag: text-classification |
|
tags: |
|
- code |
|
datasets: |
|
- shukdevdatta123/twitter_sentiment_preprocessed |
|
language: |
|
- en |
|
base_model: distilbert/distilbert-base-uncased |
|
library_name: transformers |
|
license: cc-by-nd-4.0 |
|
--- |
|
# DistilBERT-base-uncased LoRA Text Classification Model |
|
|
|
## Model Description |
|
This model is a fine-tuned version of `distilbert-base-uncased` on an unspecified dataset. It achieves the following results on the evaluation set: |
|
|
|
- **Loss:** 0.4649 |
|
- **Accuracy:** 84.16% |
|
|
|
## Intended Uses & Limitations |
|
This is a text-classification based model. |
|
|
|
## Training and Evaluation Data |
|
Look below for more details about the performances. |
|
|
|
## Steps to follow |
|
- Installing the Libraries |
|
- Loading the Dataset from HuggingFace |
|
- Train_test Split the Dataset |
|
- Model |
|
- Preprocess Data |
|
- Evaluation |
|
- Apply untrained base model("distilbert-base-uncased") to text |
|
- Train Model using LoRA |
|
- Generate Prediction |
|
- Save the Model and the Tokenizer |
|
- Load the Model and the Tokenizer to test |
|
- Push Model to HuggingFaceHub |
|
|
|
|
|
|
|
|
|
### Training Hyperparameters |
|
The following hyperparameters were used during training: |
|
|
|
- **Learning Rate:** 0.001 |
|
- **Train Batch Size:** 4 |
|
- **Eval Batch Size:** 4 |
|
- **Seed:** 42 |
|
- **Optimizer:** Adam with betas=(0.9,0.999) and epsilon=1e-08 |
|
- **LR Scheduler Type:** Linear |
|
- **Number of Epochs:** 10 |
|
|
|
### Training Results |
|
| Epoch | Training Loss | Validation Loss | Validation Accuracy | |
|
|-------|---------------|-----------------|---------------------| |
|
| 1.0 | 0.5924 | 0.5523 | 78.45% | |
|
| 2.0 | 0.5983 | 0.5236 | 80.29% | |
|
| 3.0 | 0.5703 | 0.4498 | 79.56% | |
|
| 4.0 | 0.5526 | 0.4976 | 80.66% | |
|
| 5.0 | 0.5326 | 0.4317 | 80.85% | |
|
| 6.0 | 0.5851 | 0.4562 | 82.87% | |
|
| 7.0 | 0.5466 | 0.4713 | 81.95% | |
|
| 8.0 | 0.5494 | 0.5072 | 82.50% | |
|
| 9.0 | 0.5748 | 0.4802 | 82.87% | |
|
| 10.0 | 0.5001 | 0.4649 | 84.16% | |
|
|
|
## Framework Versions |
|
- **PEFT:** 0.12.0 |
|
- **Transformers:** 4.42.4 |
|
- **PyTorch:** 2.4.0+cu121 |
|
- **Datasets:** 2.21.0 |
|
- **Tokenizers:** 0.19.1 |
|
|
|
# Dataset Viewer |
|
|
|
You can view the dataset using the following link: |
|
|
|
[View Twitter Sentiment Preprocessed Dataset](https://huggingface.co/datasets/shukdevdatta123/twitter_sentiment_preprocessed/) |
|
|
|
Simply click the link to open the dataset viewer in your browser. |
|
|
|
# Model Viewer |
|
|
|
You can view the model using the following link: |
|
|
|
[View Model in HuggingFace](https://huggingface.co/shukdevdatta123/distilbert-base-uncased-lora-text-classification/) |
|
|
|
Simply click the link to open the model file in your browser. |
|
|
|
Check out the "Fine-tune LLM.pptx" file for the theory behind this code. |
|
|
|
# Github Repository |
|
|
|
You can view the github using the following link: |
|
|
|
[View GitHub Repository](https://github.com/shukdevtroy/Fine-Tune-LLM-using-LoRA-on-custom-dataset/) |
|
|
|
Simply click the link to open the github repo in your browser. |
|
|
|
Check out the "Fine-tune LLM.pptx" file in the GitHub repo for the theory behind this code. |
|
|
|
|