distilbert-base-uncased-finetuned-imdb-mlm-acclerate
This model is a fine-tuned version of distilbert-base-uncased on the imdb dataset. It achieves the following results on the evaluation set:
- Perplexity: 11.0482
Model description
More information needed
Intended uses & limitations
More information needed
Training and evaluation data
More information needed
Training procedure
Accelerate
Training hyperparameters
The following hyperparameters were used during training:
- train_batch_size: 64
- eval_batch_size: 64
- seed: 42
- optimizer: AdamW with lr=5e-5
- lr_scheduler_type: linear
- num_epochs: 3.0
Training results
Training Loss | Epoch | Validation Loss | Perplexity |
---|---|---|---|
2.657470791203201 | 1.0 | 2.462477684020996 | 11.733848321632662 |
2.5095403741119773 | 2.0 | 2.4211950302124023 | 11.259306489817392 |
2.4733242700054388 | 3.0 | 2.4022672176361084 | 11.048196673044224 |
Run history:
Perplexity βββ
eval/loss βββββ
eval/runtime βββββ
eval/samples_per_second βββββ
eval/steps_per_second βββββ
train/epoch ββββ
ββββ
train/global_step ββββββββββββ
train/learning_rate βββ
train/loss βββ
train/total_flos β
train/train_loss β
train/train_runtime β
train/train_samples_per_second β
train/train_steps_per_second β
Run summary:
Perplexity 11.0482
eval/loss 2.41189
eval/runtime 1.923
eval/samples_per_second 520.03
eval/steps_per_second 8.32
train/epoch 3.0
train/global_step 471
train/learning_rate 0.0
train/loss 2.5354
train/total_flos 994208670720000.0
train/train_loss 2.60498
train/train_runtime 159.5259
train/train_samples_per_second 188.057
train/train_steps_per_second 2.952
View run classic-pond-2 at: https://wandb.ai/tchoud8/distilbert-base-uncased-finetuned-imdb-accelerate/runs/a7hw7i1u
Framework versions
- Transformers 4.32.0.dev0
- Pytorch 2.0.1+cu118
- Datasets 2.14.4
- Tokenizers 0.13.3
- Downloads last month
- 14
Model tree for pritam3355/distilbert-base-uncased-finetuned-imdb-mlm-accelerate
Base model
distilbert/distilbert-base-uncased