rishavranaut commited on
Commit
80976bf
·
verified ·
1 Parent(s): 5c8f699

End of training

Browse files
Files changed (2) hide show
  1. README.md +21 -20
  2. tokenizer_config.json +3 -1
README.md CHANGED
@@ -1,5 +1,6 @@
1
  ---
2
  library_name: transformers
 
3
  base_model: google/flan-t5-large
4
  tags:
5
  - generated_from_trainer
@@ -19,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
19
 
20
  This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset.
21
  It achieves the following results on the evaluation set:
22
- - Loss: 1.9154
23
- - Accuracy: 0.7753
24
- - Precision: 0.7868
25
- - Recall: 0.7553
26
- - F1 score: 0.7707
27
 
28
  ## Model description
29
 
@@ -46,7 +47,7 @@ The following hyperparameters were used during training:
46
  - train_batch_size: 1
47
  - eval_batch_size: 1
48
  - seed: 42
49
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
  - num_epochs: 5
52
 
@@ -54,22 +55,22 @@ The following hyperparameters were used during training:
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 score |
56
  |:-------------:|:------:|:-----:|:---------------:|:--------:|:---------:|:------:|:--------:|
57
- | 1.0303 | 0.4205 | 2500 | 0.8270 | 0.7471 | 0.8088 | 0.6471 | 0.7190 |
58
- | 0.9652 | 0.8410 | 5000 | 0.6861 | 0.7118 | 0.9245 | 0.4612 | 0.6154 |
59
- | 0.9232 | 1.2616 | 7500 | 0.9442 | 0.7471 | 0.8017 | 0.6565 | 0.7219 |
60
- | 0.8457 | 1.6821 | 10000 | 0.9311 | 0.7471 | 0.7869 | 0.6776 | 0.7282 |
61
- | 0.7519 | 2.1026 | 12500 | 1.0887 | 0.7682 | 0.8065 | 0.7059 | 0.7528 |
62
- | 0.6462 | 2.5231 | 15000 | 1.1780 | 0.7706 | 0.8125 | 0.7035 | 0.7541 |
63
- | 0.642 | 2.9437 | 17500 | 1.2434 | 0.7718 | 0.7954 | 0.7318 | 0.7623 |
64
- | 0.4436 | 3.3642 | 20000 | 1.3026 | 0.76 | 0.7931 | 0.7035 | 0.7456 |
65
- | 0.3762 | 3.7847 | 22500 | 1.6051 | 0.7659 | 0.7678 | 0.7624 | 0.7651 |
66
- | 0.2798 | 4.2052 | 25000 | 1.9011 | 0.7671 | 0.7683 | 0.7647 | 0.7665 |
67
- | 0.2012 | 4.6257 | 27500 | 1.9154 | 0.7753 | 0.7868 | 0.7553 | 0.7707 |
68
 
69
 
70
  ### Framework versions
71
 
72
- - Transformers 4.44.2
73
  - Pytorch 2.3.0+cu121
74
- - Datasets 2.19.1
75
- - Tokenizers 0.19.1
 
1
  ---
2
  library_name: transformers
3
+ license: apache-2.0
4
  base_model: google/flan-t5-large
5
  tags:
6
  - generated_from_trainer
 
20
 
21
  This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 2.1812
24
+ - Accuracy: 0.7706
25
+ - Precision: 0.7861
26
+ - Recall: 0.7435
27
+ - F1 score: 0.7642
28
 
29
  ## Model description
30
 
 
47
  - train_batch_size: 1
48
  - eval_batch_size: 1
49
  - seed: 42
50
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
51
  - lr_scheduler_type: linear
52
  - num_epochs: 5
53
 
 
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 score |
57
  |:-------------:|:------:|:-----:|:---------------:|:--------:|:---------:|:------:|:--------:|
58
+ | 1.1443 | 0.4205 | 2500 | 1.6635 | 0.6718 | 0.7829 | 0.4753 | 0.5915 |
59
+ | 1.0447 | 0.8410 | 5000 | 0.5585 | 0.7282 | 0.8149 | 0.5906 | 0.6849 |
60
+ | 0.9057 | 1.2616 | 7500 | 0.9051 | 0.7318 | 0.7275 | 0.7412 | 0.7343 |
61
+ | 0.8348 | 1.6821 | 10000 | 0.6307 | 0.7659 | 0.8742 | 0.6212 | 0.7263 |
62
+ | 0.7331 | 2.1026 | 12500 | 0.9500 | 0.7612 | 0.7489 | 0.7859 | 0.7669 |
63
+ | 0.6167 | 2.5231 | 15000 | 1.1524 | 0.7788 | 0.7970 | 0.7482 | 0.7718 |
64
+ | 0.6209 | 2.9437 | 17500 | 1.1690 | 0.7635 | 0.7872 | 0.7224 | 0.7534 |
65
+ | 0.4411 | 3.3642 | 20000 | 1.7563 | 0.7847 | 0.8438 | 0.6988 | 0.7645 |
66
+ | 0.4196 | 3.7847 | 22500 | 1.7767 | 0.7412 | 0.7204 | 0.7882 | 0.7528 |
67
+ | 0.292 | 4.2052 | 25000 | 2.0410 | 0.7624 | 0.7648 | 0.7576 | 0.7612 |
68
+ | 0.1791 | 4.6257 | 27500 | 2.1812 | 0.7706 | 0.7861 | 0.7435 | 0.7642 |
69
 
70
 
71
  ### Framework versions
72
 
73
+ - Transformers 4.48.3
74
  - Pytorch 2.3.0+cu121
75
+ - Datasets 3.2.0
76
+ - Tokenizers 0.21.0
tokenizer_config.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<pad>",
@@ -927,9 +928,10 @@
927
  "<extra_id_98>",
928
  "<extra_id_99>"
929
  ],
930
- "clean_up_tokenization_spaces": true,
931
  "eos_token": "</s>",
932
  "extra_ids": 100,
 
933
  "model_max_length": 512,
934
  "pad_token": "<pad>",
935
  "sp_model_kwargs": {},
 
1
  {
2
+ "add_prefix_space": null,
3
  "added_tokens_decoder": {
4
  "0": {
5
  "content": "<pad>",
 
928
  "<extra_id_98>",
929
  "<extra_id_99>"
930
  ],
931
+ "clean_up_tokenization_spaces": false,
932
  "eos_token": "</s>",
933
  "extra_ids": 100,
934
+ "extra_special_tokens": {},
935
  "model_max_length": 512,
936
  "pad_token": "<pad>",
937
  "sp_model_kwargs": {},