shorecode commited on
Commit
ee69ad7
·
verified ·
1 Parent(s): f7984bd

Training complete!

Browse files
Files changed (3) hide show
  1. README.md +8 -10
  2. model.safetensors +1 -1
  3. training_args.bin +1 -1
README.md CHANGED
@@ -16,7 +16,12 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  This model is a fine-tuned version of [google/t5-efficient-tiny-nh8](https://huggingface.co/google/t5-efficient-tiny-nh8) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 3.8741
 
 
 
 
 
20
 
21
  ## Model description
22
 
@@ -36,21 +41,14 @@ More information needed
36
 
37
  The following hyperparameters were used during training:
38
  - learning_rate: 7.000000000000001e-05
39
- - train_batch_size: 50
40
- - eval_batch_size: 50
41
  - seed: 42
42
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
43
  - lr_scheduler_type: linear
44
  - num_epochs: 3
45
  - mixed_precision_training: Native AMP
46
 
47
- ### Training results
48
-
49
- | Training Loss | Epoch | Step | Validation Loss |
50
- |:-------------:|:-----:|:----:|:---------------:|
51
- | 4.3365 | 2.5 | 200 | 3.8741 |
52
-
53
-
54
  ### Framework versions
55
 
56
  - Transformers 4.47.0
 
16
 
17
  This model is a fine-tuned version of [google/t5-efficient-tiny-nh8](https://huggingface.co/google/t5-efficient-tiny-nh8) on an unknown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - eval_loss: 4.7168
20
+ - eval_model_preparation_time: 0.0032
21
+ - eval_runtime: 23.0401
22
+ - eval_samples_per_second: 347.612
23
+ - eval_steps_per_second: 4.991
24
+ - step: 0
25
 
26
  ## Model description
27
 
 
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 7.000000000000001e-05
44
+ - train_batch_size: 70
45
+ - eval_batch_size: 70
46
  - seed: 42
47
  - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - num_epochs: 3
50
  - mixed_precision_training: Native AMP
51
 
 
 
 
 
 
 
 
52
  ### Framework versions
53
 
54
  - Transformers 4.47.0
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:32d2c74aa61047ec79b33a0624c061d736c60f81f92dd06bec96522458271259
3
  size 62293080
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a2de4de5443fa966efc402c8dfda08d1bedfae86be35b1f616a60ad28e7a990
3
  size 62293080
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f721dd7c177c395969980c852b91bcf9a8013cd375af2d1a26b3ec20012eb08f
3
  size 5304
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:639965cd1c0265d0ae1ef8aafaa8aec94659deccc7dbe06d4dc3452d468701c7
3
  size 5304