End of training
Browse files- README.md +21 -20
- tokenizer_config.json +3 -1
README.md
CHANGED
@@ -1,5 +1,6 @@
|
|
1 |
---
|
2 |
library_name: transformers
|
|
|
3 |
base_model: google/flan-t5-large
|
4 |
tags:
|
5 |
- generated_from_trainer
|
@@ -19,11 +20,11 @@ should probably proofread and complete it, then remove this comment. -->
|
|
19 |
|
20 |
This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset.
|
21 |
It achieves the following results on the evaluation set:
|
22 |
-
- Loss:
|
23 |
-
- Accuracy: 0.
|
24 |
-
- Precision: 0.
|
25 |
-
- Recall: 0.
|
26 |
-
- F1 score: 0.
|
27 |
|
28 |
## Model description
|
29 |
|
@@ -46,7 +47,7 @@ The following hyperparameters were used during training:
|
|
46 |
- train_batch_size: 1
|
47 |
- eval_batch_size: 1
|
48 |
- seed: 42
|
49 |
-
- optimizer:
|
50 |
- lr_scheduler_type: linear
|
51 |
- num_epochs: 5
|
52 |
|
@@ -54,22 +55,22 @@ The following hyperparameters were used during training:
|
|
54 |
|
55 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 score |
|
56 |
|:-------------:|:------:|:-----:|:---------------:|:--------:|:---------:|:------:|:--------:|
|
57 |
-
| 1.
|
58 |
-
|
|
59 |
-
| 0.
|
60 |
-
| 0.
|
61 |
-
| 0.
|
62 |
-
| 0.
|
63 |
-
| 0.
|
64 |
-
| 0.
|
65 |
-
| 0.
|
66 |
-
| 0.
|
67 |
-
| 0.
|
68 |
|
69 |
|
70 |
### Framework versions
|
71 |
|
72 |
-
- Transformers 4.
|
73 |
- Pytorch 2.3.0+cu121
|
74 |
-
- Datasets 2.
|
75 |
-
- Tokenizers 0.
|
|
|
1 |
---
|
2 |
library_name: transformers
|
3 |
+
license: apache-2.0
|
4 |
base_model: google/flan-t5-large
|
5 |
tags:
|
6 |
- generated_from_trainer
|
|
|
20 |
|
21 |
This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on the None dataset.
|
22 |
It achieves the following results on the evaluation set:
|
23 |
+
- Loss: 2.1812
|
24 |
+
- Accuracy: 0.7706
|
25 |
+
- Precision: 0.7861
|
26 |
+
- Recall: 0.7435
|
27 |
+
- F1 score: 0.7642
|
28 |
|
29 |
## Model description
|
30 |
|
|
|
47 |
- train_batch_size: 1
|
48 |
- eval_batch_size: 1
|
49 |
- seed: 42
|
50 |
+
- optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
|
51 |
- lr_scheduler_type: linear
|
52 |
- num_epochs: 5
|
53 |
|
|
|
55 |
|
56 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy | Precision | Recall | F1 score |
|
57 |
|:-------------:|:------:|:-----:|:---------------:|:--------:|:---------:|:------:|:--------:|
|
58 |
+
| 1.1443 | 0.4205 | 2500 | 1.6635 | 0.6718 | 0.7829 | 0.4753 | 0.5915 |
|
59 |
+
| 1.0447 | 0.8410 | 5000 | 0.5585 | 0.7282 | 0.8149 | 0.5906 | 0.6849 |
|
60 |
+
| 0.9057 | 1.2616 | 7500 | 0.9051 | 0.7318 | 0.7275 | 0.7412 | 0.7343 |
|
61 |
+
| 0.8348 | 1.6821 | 10000 | 0.6307 | 0.7659 | 0.8742 | 0.6212 | 0.7263 |
|
62 |
+
| 0.7331 | 2.1026 | 12500 | 0.9500 | 0.7612 | 0.7489 | 0.7859 | 0.7669 |
|
63 |
+
| 0.6167 | 2.5231 | 15000 | 1.1524 | 0.7788 | 0.7970 | 0.7482 | 0.7718 |
|
64 |
+
| 0.6209 | 2.9437 | 17500 | 1.1690 | 0.7635 | 0.7872 | 0.7224 | 0.7534 |
|
65 |
+
| 0.4411 | 3.3642 | 20000 | 1.7563 | 0.7847 | 0.8438 | 0.6988 | 0.7645 |
|
66 |
+
| 0.4196 | 3.7847 | 22500 | 1.7767 | 0.7412 | 0.7204 | 0.7882 | 0.7528 |
|
67 |
+
| 0.292 | 4.2052 | 25000 | 2.0410 | 0.7624 | 0.7648 | 0.7576 | 0.7612 |
|
68 |
+
| 0.1791 | 4.6257 | 27500 | 2.1812 | 0.7706 | 0.7861 | 0.7435 | 0.7642 |
|
69 |
|
70 |
|
71 |
### Framework versions
|
72 |
|
73 |
+
- Transformers 4.48.3
|
74 |
- Pytorch 2.3.0+cu121
|
75 |
+
- Datasets 3.2.0
|
76 |
+
- Tokenizers 0.21.0
|
tokenizer_config.json
CHANGED
@@ -1,4 +1,5 @@
|
|
1 |
{
|
|
|
2 |
"added_tokens_decoder": {
|
3 |
"0": {
|
4 |
"content": "<pad>",
|
@@ -927,9 +928,10 @@
|
|
927 |
"<extra_id_98>",
|
928 |
"<extra_id_99>"
|
929 |
],
|
930 |
-
"clean_up_tokenization_spaces":
|
931 |
"eos_token": "</s>",
|
932 |
"extra_ids": 100,
|
|
|
933 |
"model_max_length": 512,
|
934 |
"pad_token": "<pad>",
|
935 |
"sp_model_kwargs": {},
|
|
|
1 |
{
|
2 |
+
"add_prefix_space": null,
|
3 |
"added_tokens_decoder": {
|
4 |
"0": {
|
5 |
"content": "<pad>",
|
|
|
928 |
"<extra_id_98>",
|
929 |
"<extra_id_99>"
|
930 |
],
|
931 |
+
"clean_up_tokenization_spaces": false,
|
932 |
"eos_token": "</s>",
|
933 |
"extra_ids": 100,
|
934 |
+
"extra_special_tokens": {},
|
935 |
"model_max_length": 512,
|
936 |
"pad_token": "<pad>",
|
937 |
"sp_model_kwargs": {},
|