navidmadani commited on
Commit
f796d97
·
verified ·
1 Parent(s): c7194c9

Upload 13 files

Browse files
README.md CHANGED
@@ -1,3 +1,85 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: google/flan-t5-large
4
+ tags:
5
+ - generated_from_trainer
6
+ metrics:
7
+ - rouge
8
+ model-index:
9
+ - name: deductor-flant5-large
10
+ results: []
11
+ ---
12
+
13
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
+ should probably proofread and complete it, then remove this comment. -->
15
+
16
+ # deductor-flant5-large
17
+
18
+ This model is a fine-tuned version of [google/flan-t5-large](https://huggingface.co/google/flan-t5-large) on an unknown dataset.
19
+ It achieves the following results on the evaluation set:
20
+ - Loss: 0.2461
21
+ - Rouge1: 92.1213
22
+ - Rouge2: 86.4281
23
+ - Rougel: 90.5846
24
+ - Rougelsum: 90.5294
25
+ - Gen Len: 11.2014
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 5e-05
45
+ - train_batch_size: 16
46
+ - eval_batch_size: 32
47
+ - seed: 42
48
+ - gradient_accumulation_steps: 4
49
+ - total_train_batch_size: 64
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - num_epochs: 10.0
53
+
54
+ ### Training results
55
+
56
+ | Training Loss | Epoch | Step | Validation Loss | Rouge1 | Rouge2 | Rougel | Rougelsum | Gen Len |
57
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|:-------:|:-------:|:---------:|:-------:|
58
+ | 0.306 | 0.19 | 50 | 0.2959 | 89.3028 | 82.5127 | 87.4173 | 87.3544 | 11.2211 |
59
+ | 0.2774 | 0.38 | 100 | 0.2717 | 90.8414 | 84.2378 | 88.9385 | 88.9058 | 11.2571 |
60
+ | 0.2366 | 0.57 | 150 | 0.2613 | 91.0152 | 84.6687 | 89.2107 | 89.1735 | 11.2081 |
61
+ | 0.2166 | 0.77 | 200 | 0.2585 | 91.5215 | 85.4308 | 89.7742 | 89.7422 | 11.2802 |
62
+ | 0.22 | 0.96 | 250 | 0.2517 | 91.5587 | 85.6107 | 89.8835 | 89.8621 | 11.2655 |
63
+ | 0.1564 | 1.15 | 300 | 0.2630 | 91.999 | 86.0835 | 90.3611 | 90.3168 | 11.2039 |
64
+ | 0.1803 | 1.34 | 350 | 0.2546 | 91.5183 | 85.6214 | 89.9752 | 89.9323 | 11.2462 |
65
+ | 0.1737 | 1.53 | 400 | 0.2483 | 91.8342 | 86.0171 | 90.3042 | 90.2641 | 11.1943 |
66
+ | 0.157 | 1.72 | 450 | 0.2493 | 91.6585 | 85.4651 | 90.0181 | 89.9991 | 10.9376 |
67
+ | 0.1561 | 1.92 | 500 | 0.2461 | 92.1213 | 86.4281 | 90.5846 | 90.5294 | 11.2014 |
68
+ | 0.1191 | 2.11 | 550 | 0.2585 | 92.4493 | 86.6961 | 90.9293 | 90.8761 | 11.2416 |
69
+ | 0.1134 | 2.3 | 600 | 0.2633 | 92.4707 | 86.833 | 90.9516 | 90.9195 | 11.1675 |
70
+ | 0.1227 | 2.49 | 650 | 0.2592 | 92.2738 | 86.5064 | 90.7556 | 90.6998 | 11.2642 |
71
+ | 0.1175 | 2.68 | 700 | 0.2657 | 92.0861 | 86.2203 | 90.6168 | 90.5657 | 11.1700 |
72
+ | 0.1132 | 2.87 | 750 | 0.2644 | 92.3834 | 86.7237 | 90.8761 | 90.8389 | 11.2123 |
73
+ | 0.1097 | 3.07 | 800 | 0.2692 | 92.3356 | 86.7021 | 90.8717 | 90.8185 | 11.1822 |
74
+ | 0.0949 | 3.26 | 850 | 0.2690 | 92.5746 | 87.001 | 91.1734 | 91.1222 | 11.2785 |
75
+ | 0.0813 | 3.45 | 900 | 0.2875 | 92.5641 | 86.9813 | 91.0881 | 91.0411 | 11.2257 |
76
+ | 0.0861 | 3.64 | 950 | 0.2800 | 92.4738 | 86.9379 | 91.0384 | 90.9995 | 11.2136 |
77
+ | 0.0879 | 3.83 | 1000 | 0.2770 | 92.6025 | 87.105 | 91.1632 | 91.1292 | 11.2303 |
78
+
79
+
80
+ ### Framework versions
81
+
82
+ - Transformers 4.36.2
83
+ - Pytorch 2.0.1
84
+ - Datasets 2.18.0
85
+ - Tokenizers 0.15.2
all_results.json ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.83,
3
+ "eval_gen_len": 11.20142378559464,
4
+ "eval_loss": 0.24614077806472778,
5
+ "eval_rouge1": 92.1213,
6
+ "eval_rouge2": 86.4281,
7
+ "eval_rougeL": 90.5846,
8
+ "eval_rougeLsum": 90.5294,
9
+ "eval_runtime": 58.4188,
10
+ "eval_samples": 2388,
11
+ "eval_samples_per_second": 40.877,
12
+ "eval_steps_per_second": 1.284,
13
+ "train_loss": 0.16499798774719238,
14
+ "train_runtime": 3538.4305,
15
+ "train_samples": 16700,
16
+ "train_samples_per_second": 47.196,
17
+ "train_steps_per_second": 0.738
18
+ }
config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/flan-t5-large",
3
+ "architectures": [
4
+ "T5ForConditionalGeneration"
5
+ ],
6
+ "classifier_dropout": 0.0,
7
+ "d_ff": 2816,
8
+ "d_kv": 64,
9
+ "d_model": 1024,
10
+ "decoder_start_token_id": 0,
11
+ "dense_act_fn": "gelu_new",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 1,
14
+ "feed_forward_proj": "gated-gelu",
15
+ "initializer_factor": 1.0,
16
+ "is_encoder_decoder": true,
17
+ "is_gated_act": true,
18
+ "layer_norm_epsilon": 1e-06,
19
+ "model_type": "t5",
20
+ "n_positions": 512,
21
+ "num_decoder_layers": 24,
22
+ "num_heads": 16,
23
+ "num_layers": 24,
24
+ "output_past": true,
25
+ "pad_token_id": 0,
26
+ "relative_attention_max_distance": 128,
27
+ "relative_attention_num_buckets": 32,
28
+ "tie_word_embeddings": false,
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.36.2",
31
+ "use_cache": true,
32
+ "vocab_size": 32128
33
+ }
eval_results.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.83,
3
+ "eval_gen_len": 11.20142378559464,
4
+ "eval_loss": 0.24614077806472778,
5
+ "eval_rouge1": 92.1213,
6
+ "eval_rouge2": 86.4281,
7
+ "eval_rougeL": 90.5846,
8
+ "eval_rougeLsum": 90.5294,
9
+ "eval_runtime": 58.4188,
10
+ "eval_samples": 2388,
11
+ "eval_samples_per_second": 40.877,
12
+ "eval_steps_per_second": 1.284
13
+ }
generation_config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "max_length": 256,
5
+ "pad_token_id": 0,
6
+ "transformers_version": "4.36.2"
7
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4209f37f11e36519483138a54b22bf63c706fb8840cca15534ba1c5ec01d9d26
3
+ size 3132668808
special_tokens_map.json ADDED
@@ -0,0 +1,125 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<extra_id_0>",
4
+ "<extra_id_1>",
5
+ "<extra_id_2>",
6
+ "<extra_id_3>",
7
+ "<extra_id_4>",
8
+ "<extra_id_5>",
9
+ "<extra_id_6>",
10
+ "<extra_id_7>",
11
+ "<extra_id_8>",
12
+ "<extra_id_9>",
13
+ "<extra_id_10>",
14
+ "<extra_id_11>",
15
+ "<extra_id_12>",
16
+ "<extra_id_13>",
17
+ "<extra_id_14>",
18
+ "<extra_id_15>",
19
+ "<extra_id_16>",
20
+ "<extra_id_17>",
21
+ "<extra_id_18>",
22
+ "<extra_id_19>",
23
+ "<extra_id_20>",
24
+ "<extra_id_21>",
25
+ "<extra_id_22>",
26
+ "<extra_id_23>",
27
+ "<extra_id_24>",
28
+ "<extra_id_25>",
29
+ "<extra_id_26>",
30
+ "<extra_id_27>",
31
+ "<extra_id_28>",
32
+ "<extra_id_29>",
33
+ "<extra_id_30>",
34
+ "<extra_id_31>",
35
+ "<extra_id_32>",
36
+ "<extra_id_33>",
37
+ "<extra_id_34>",
38
+ "<extra_id_35>",
39
+ "<extra_id_36>",
40
+ "<extra_id_37>",
41
+ "<extra_id_38>",
42
+ "<extra_id_39>",
43
+ "<extra_id_40>",
44
+ "<extra_id_41>",
45
+ "<extra_id_42>",
46
+ "<extra_id_43>",
47
+ "<extra_id_44>",
48
+ "<extra_id_45>",
49
+ "<extra_id_46>",
50
+ "<extra_id_47>",
51
+ "<extra_id_48>",
52
+ "<extra_id_49>",
53
+ "<extra_id_50>",
54
+ "<extra_id_51>",
55
+ "<extra_id_52>",
56
+ "<extra_id_53>",
57
+ "<extra_id_54>",
58
+ "<extra_id_55>",
59
+ "<extra_id_56>",
60
+ "<extra_id_57>",
61
+ "<extra_id_58>",
62
+ "<extra_id_59>",
63
+ "<extra_id_60>",
64
+ "<extra_id_61>",
65
+ "<extra_id_62>",
66
+ "<extra_id_63>",
67
+ "<extra_id_64>",
68
+ "<extra_id_65>",
69
+ "<extra_id_66>",
70
+ "<extra_id_67>",
71
+ "<extra_id_68>",
72
+ "<extra_id_69>",
73
+ "<extra_id_70>",
74
+ "<extra_id_71>",
75
+ "<extra_id_72>",
76
+ "<extra_id_73>",
77
+ "<extra_id_74>",
78
+ "<extra_id_75>",
79
+ "<extra_id_76>",
80
+ "<extra_id_77>",
81
+ "<extra_id_78>",
82
+ "<extra_id_79>",
83
+ "<extra_id_80>",
84
+ "<extra_id_81>",
85
+ "<extra_id_82>",
86
+ "<extra_id_83>",
87
+ "<extra_id_84>",
88
+ "<extra_id_85>",
89
+ "<extra_id_86>",
90
+ "<extra_id_87>",
91
+ "<extra_id_88>",
92
+ "<extra_id_89>",
93
+ "<extra_id_90>",
94
+ "<extra_id_91>",
95
+ "<extra_id_92>",
96
+ "<extra_id_93>",
97
+ "<extra_id_94>",
98
+ "<extra_id_95>",
99
+ "<extra_id_96>",
100
+ "<extra_id_97>",
101
+ "<extra_id_98>",
102
+ "<extra_id_99>"
103
+ ],
104
+ "eos_token": {
105
+ "content": "</s>",
106
+ "lstrip": false,
107
+ "normalized": false,
108
+ "rstrip": false,
109
+ "single_word": false
110
+ },
111
+ "pad_token": {
112
+ "content": "<pad>",
113
+ "lstrip": false,
114
+ "normalized": false,
115
+ "rstrip": false,
116
+ "single_word": false
117
+ },
118
+ "unk_token": {
119
+ "content": "<unk>",
120
+ "lstrip": false,
121
+ "normalized": false,
122
+ "rstrip": false,
123
+ "single_word": false
124
+ }
125
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d60acb128cf7b7f2536e8f38a5b18a05535c9e14c7a355904270e15b0945ea86
3
+ size 791656
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,938 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<pad>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "</s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<unk>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "32000": {
28
+ "content": "<extra_id_99>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "32001": {
36
+ "content": "<extra_id_98>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "32002": {
44
+ "content": "<extra_id_97>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "32003": {
52
+ "content": "<extra_id_96>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "32004": {
60
+ "content": "<extra_id_95>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "32005": {
68
+ "content": "<extra_id_94>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "32006": {
76
+ "content": "<extra_id_93>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "32007": {
84
+ "content": "<extra_id_92>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "32008": {
92
+ "content": "<extra_id_91>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "32009": {
100
+ "content": "<extra_id_90>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "32010": {
108
+ "content": "<extra_id_89>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "32011": {
116
+ "content": "<extra_id_88>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "32012": {
124
+ "content": "<extra_id_87>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "32013": {
132
+ "content": "<extra_id_86>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "32014": {
140
+ "content": "<extra_id_85>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "32015": {
148
+ "content": "<extra_id_84>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "32016": {
156
+ "content": "<extra_id_83>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "32017": {
164
+ "content": "<extra_id_82>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "32018": {
172
+ "content": "<extra_id_81>",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "32019": {
180
+ "content": "<extra_id_80>",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "32020": {
188
+ "content": "<extra_id_79>",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "32021": {
196
+ "content": "<extra_id_78>",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "32022": {
204
+ "content": "<extra_id_77>",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "32023": {
212
+ "content": "<extra_id_76>",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "32024": {
220
+ "content": "<extra_id_75>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "32025": {
228
+ "content": "<extra_id_74>",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "32026": {
236
+ "content": "<extra_id_73>",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "32027": {
244
+ "content": "<extra_id_72>",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "32028": {
252
+ "content": "<extra_id_71>",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "32029": {
260
+ "content": "<extra_id_70>",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "32030": {
268
+ "content": "<extra_id_69>",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "32031": {
276
+ "content": "<extra_id_68>",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "32032": {
284
+ "content": "<extra_id_67>",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "32033": {
292
+ "content": "<extra_id_66>",
293
+ "lstrip": false,
294
+ "normalized": false,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "32034": {
300
+ "content": "<extra_id_65>",
301
+ "lstrip": false,
302
+ "normalized": false,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "32035": {
308
+ "content": "<extra_id_64>",
309
+ "lstrip": false,
310
+ "normalized": false,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "32036": {
316
+ "content": "<extra_id_63>",
317
+ "lstrip": false,
318
+ "normalized": false,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "32037": {
324
+ "content": "<extra_id_62>",
325
+ "lstrip": false,
326
+ "normalized": false,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "32038": {
332
+ "content": "<extra_id_61>",
333
+ "lstrip": false,
334
+ "normalized": false,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "32039": {
340
+ "content": "<extra_id_60>",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "32040": {
348
+ "content": "<extra_id_59>",
349
+ "lstrip": false,
350
+ "normalized": false,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "32041": {
356
+ "content": "<extra_id_58>",
357
+ "lstrip": false,
358
+ "normalized": false,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "32042": {
364
+ "content": "<extra_id_57>",
365
+ "lstrip": false,
366
+ "normalized": false,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "32043": {
372
+ "content": "<extra_id_56>",
373
+ "lstrip": false,
374
+ "normalized": false,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "32044": {
380
+ "content": "<extra_id_55>",
381
+ "lstrip": false,
382
+ "normalized": false,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": true
386
+ },
387
+ "32045": {
388
+ "content": "<extra_id_54>",
389
+ "lstrip": false,
390
+ "normalized": false,
391
+ "rstrip": false,
392
+ "single_word": false,
393
+ "special": true
394
+ },
395
+ "32046": {
396
+ "content": "<extra_id_53>",
397
+ "lstrip": false,
398
+ "normalized": false,
399
+ "rstrip": false,
400
+ "single_word": false,
401
+ "special": true
402
+ },
403
+ "32047": {
404
+ "content": "<extra_id_52>",
405
+ "lstrip": false,
406
+ "normalized": false,
407
+ "rstrip": false,
408
+ "single_word": false,
409
+ "special": true
410
+ },
411
+ "32048": {
412
+ "content": "<extra_id_51>",
413
+ "lstrip": false,
414
+ "normalized": false,
415
+ "rstrip": false,
416
+ "single_word": false,
417
+ "special": true
418
+ },
419
+ "32049": {
420
+ "content": "<extra_id_50>",
421
+ "lstrip": false,
422
+ "normalized": false,
423
+ "rstrip": false,
424
+ "single_word": false,
425
+ "special": true
426
+ },
427
+ "32050": {
428
+ "content": "<extra_id_49>",
429
+ "lstrip": false,
430
+ "normalized": false,
431
+ "rstrip": false,
432
+ "single_word": false,
433
+ "special": true
434
+ },
435
+ "32051": {
436
+ "content": "<extra_id_48>",
437
+ "lstrip": false,
438
+ "normalized": false,
439
+ "rstrip": false,
440
+ "single_word": false,
441
+ "special": true
442
+ },
443
+ "32052": {
444
+ "content": "<extra_id_47>",
445
+ "lstrip": false,
446
+ "normalized": false,
447
+ "rstrip": false,
448
+ "single_word": false,
449
+ "special": true
450
+ },
451
+ "32053": {
452
+ "content": "<extra_id_46>",
453
+ "lstrip": false,
454
+ "normalized": false,
455
+ "rstrip": false,
456
+ "single_word": false,
457
+ "special": true
458
+ },
459
+ "32054": {
460
+ "content": "<extra_id_45>",
461
+ "lstrip": false,
462
+ "normalized": false,
463
+ "rstrip": false,
464
+ "single_word": false,
465
+ "special": true
466
+ },
467
+ "32055": {
468
+ "content": "<extra_id_44>",
469
+ "lstrip": false,
470
+ "normalized": false,
471
+ "rstrip": false,
472
+ "single_word": false,
473
+ "special": true
474
+ },
475
+ "32056": {
476
+ "content": "<extra_id_43>",
477
+ "lstrip": false,
478
+ "normalized": false,
479
+ "rstrip": false,
480
+ "single_word": false,
481
+ "special": true
482
+ },
483
+ "32057": {
484
+ "content": "<extra_id_42>",
485
+ "lstrip": false,
486
+ "normalized": false,
487
+ "rstrip": false,
488
+ "single_word": false,
489
+ "special": true
490
+ },
491
+ "32058": {
492
+ "content": "<extra_id_41>",
493
+ "lstrip": false,
494
+ "normalized": false,
495
+ "rstrip": false,
496
+ "single_word": false,
497
+ "special": true
498
+ },
499
+ "32059": {
500
+ "content": "<extra_id_40>",
501
+ "lstrip": false,
502
+ "normalized": false,
503
+ "rstrip": false,
504
+ "single_word": false,
505
+ "special": true
506
+ },
507
+ "32060": {
508
+ "content": "<extra_id_39>",
509
+ "lstrip": false,
510
+ "normalized": false,
511
+ "rstrip": false,
512
+ "single_word": false,
513
+ "special": true
514
+ },
515
+ "32061": {
516
+ "content": "<extra_id_38>",
517
+ "lstrip": false,
518
+ "normalized": false,
519
+ "rstrip": false,
520
+ "single_word": false,
521
+ "special": true
522
+ },
523
+ "32062": {
524
+ "content": "<extra_id_37>",
525
+ "lstrip": false,
526
+ "normalized": false,
527
+ "rstrip": false,
528
+ "single_word": false,
529
+ "special": true
530
+ },
531
+ "32063": {
532
+ "content": "<extra_id_36>",
533
+ "lstrip": false,
534
+ "normalized": false,
535
+ "rstrip": false,
536
+ "single_word": false,
537
+ "special": true
538
+ },
539
+ "32064": {
540
+ "content": "<extra_id_35>",
541
+ "lstrip": false,
542
+ "normalized": false,
543
+ "rstrip": false,
544
+ "single_word": false,
545
+ "special": true
546
+ },
547
+ "32065": {
548
+ "content": "<extra_id_34>",
549
+ "lstrip": false,
550
+ "normalized": false,
551
+ "rstrip": false,
552
+ "single_word": false,
553
+ "special": true
554
+ },
555
+ "32066": {
556
+ "content": "<extra_id_33>",
557
+ "lstrip": false,
558
+ "normalized": false,
559
+ "rstrip": false,
560
+ "single_word": false,
561
+ "special": true
562
+ },
563
+ "32067": {
564
+ "content": "<extra_id_32>",
565
+ "lstrip": false,
566
+ "normalized": false,
567
+ "rstrip": false,
568
+ "single_word": false,
569
+ "special": true
570
+ },
571
+ "32068": {
572
+ "content": "<extra_id_31>",
573
+ "lstrip": false,
574
+ "normalized": false,
575
+ "rstrip": false,
576
+ "single_word": false,
577
+ "special": true
578
+ },
579
+ "32069": {
580
+ "content": "<extra_id_30>",
581
+ "lstrip": false,
582
+ "normalized": false,
583
+ "rstrip": false,
584
+ "single_word": false,
585
+ "special": true
586
+ },
587
+ "32070": {
588
+ "content": "<extra_id_29>",
589
+ "lstrip": false,
590
+ "normalized": false,
591
+ "rstrip": false,
592
+ "single_word": false,
593
+ "special": true
594
+ },
595
+ "32071": {
596
+ "content": "<extra_id_28>",
597
+ "lstrip": false,
598
+ "normalized": false,
599
+ "rstrip": false,
600
+ "single_word": false,
601
+ "special": true
602
+ },
603
+ "32072": {
604
+ "content": "<extra_id_27>",
605
+ "lstrip": false,
606
+ "normalized": false,
607
+ "rstrip": false,
608
+ "single_word": false,
609
+ "special": true
610
+ },
611
+ "32073": {
612
+ "content": "<extra_id_26>",
613
+ "lstrip": false,
614
+ "normalized": false,
615
+ "rstrip": false,
616
+ "single_word": false,
617
+ "special": true
618
+ },
619
+ "32074": {
620
+ "content": "<extra_id_25>",
621
+ "lstrip": false,
622
+ "normalized": false,
623
+ "rstrip": false,
624
+ "single_word": false,
625
+ "special": true
626
+ },
627
+ "32075": {
628
+ "content": "<extra_id_24>",
629
+ "lstrip": false,
630
+ "normalized": false,
631
+ "rstrip": false,
632
+ "single_word": false,
633
+ "special": true
634
+ },
635
+ "32076": {
636
+ "content": "<extra_id_23>",
637
+ "lstrip": false,
638
+ "normalized": false,
639
+ "rstrip": false,
640
+ "single_word": false,
641
+ "special": true
642
+ },
643
+ "32077": {
644
+ "content": "<extra_id_22>",
645
+ "lstrip": false,
646
+ "normalized": false,
647
+ "rstrip": false,
648
+ "single_word": false,
649
+ "special": true
650
+ },
651
+ "32078": {
652
+ "content": "<extra_id_21>",
653
+ "lstrip": false,
654
+ "normalized": false,
655
+ "rstrip": false,
656
+ "single_word": false,
657
+ "special": true
658
+ },
659
+ "32079": {
660
+ "content": "<extra_id_20>",
661
+ "lstrip": false,
662
+ "normalized": false,
663
+ "rstrip": false,
664
+ "single_word": false,
665
+ "special": true
666
+ },
667
+ "32080": {
668
+ "content": "<extra_id_19>",
669
+ "lstrip": false,
670
+ "normalized": false,
671
+ "rstrip": false,
672
+ "single_word": false,
673
+ "special": true
674
+ },
675
+ "32081": {
676
+ "content": "<extra_id_18>",
677
+ "lstrip": false,
678
+ "normalized": false,
679
+ "rstrip": false,
680
+ "single_word": false,
681
+ "special": true
682
+ },
683
+ "32082": {
684
+ "content": "<extra_id_17>",
685
+ "lstrip": false,
686
+ "normalized": false,
687
+ "rstrip": false,
688
+ "single_word": false,
689
+ "special": true
690
+ },
691
+ "32083": {
692
+ "content": "<extra_id_16>",
693
+ "lstrip": false,
694
+ "normalized": false,
695
+ "rstrip": false,
696
+ "single_word": false,
697
+ "special": true
698
+ },
699
+ "32084": {
700
+ "content": "<extra_id_15>",
701
+ "lstrip": false,
702
+ "normalized": false,
703
+ "rstrip": false,
704
+ "single_word": false,
705
+ "special": true
706
+ },
707
+ "32085": {
708
+ "content": "<extra_id_14>",
709
+ "lstrip": false,
710
+ "normalized": false,
711
+ "rstrip": false,
712
+ "single_word": false,
713
+ "special": true
714
+ },
715
+ "32086": {
716
+ "content": "<extra_id_13>",
717
+ "lstrip": false,
718
+ "normalized": false,
719
+ "rstrip": false,
720
+ "single_word": false,
721
+ "special": true
722
+ },
723
+ "32087": {
724
+ "content": "<extra_id_12>",
725
+ "lstrip": false,
726
+ "normalized": false,
727
+ "rstrip": false,
728
+ "single_word": false,
729
+ "special": true
730
+ },
731
+ "32088": {
732
+ "content": "<extra_id_11>",
733
+ "lstrip": false,
734
+ "normalized": false,
735
+ "rstrip": false,
736
+ "single_word": false,
737
+ "special": true
738
+ },
739
+ "32089": {
740
+ "content": "<extra_id_10>",
741
+ "lstrip": false,
742
+ "normalized": false,
743
+ "rstrip": false,
744
+ "single_word": false,
745
+ "special": true
746
+ },
747
+ "32090": {
748
+ "content": "<extra_id_9>",
749
+ "lstrip": false,
750
+ "normalized": false,
751
+ "rstrip": false,
752
+ "single_word": false,
753
+ "special": true
754
+ },
755
+ "32091": {
756
+ "content": "<extra_id_8>",
757
+ "lstrip": false,
758
+ "normalized": false,
759
+ "rstrip": false,
760
+ "single_word": false,
761
+ "special": true
762
+ },
763
+ "32092": {
764
+ "content": "<extra_id_7>",
765
+ "lstrip": false,
766
+ "normalized": false,
767
+ "rstrip": false,
768
+ "single_word": false,
769
+ "special": true
770
+ },
771
+ "32093": {
772
+ "content": "<extra_id_6>",
773
+ "lstrip": false,
774
+ "normalized": false,
775
+ "rstrip": false,
776
+ "single_word": false,
777
+ "special": true
778
+ },
779
+ "32094": {
780
+ "content": "<extra_id_5>",
781
+ "lstrip": false,
782
+ "normalized": false,
783
+ "rstrip": false,
784
+ "single_word": false,
785
+ "special": true
786
+ },
787
+ "32095": {
788
+ "content": "<extra_id_4>",
789
+ "lstrip": false,
790
+ "normalized": false,
791
+ "rstrip": false,
792
+ "single_word": false,
793
+ "special": true
794
+ },
795
+ "32096": {
796
+ "content": "<extra_id_3>",
797
+ "lstrip": false,
798
+ "normalized": false,
799
+ "rstrip": false,
800
+ "single_word": false,
801
+ "special": true
802
+ },
803
+ "32097": {
804
+ "content": "<extra_id_2>",
805
+ "lstrip": false,
806
+ "normalized": false,
807
+ "rstrip": false,
808
+ "single_word": false,
809
+ "special": true
810
+ },
811
+ "32098": {
812
+ "content": "<extra_id_1>",
813
+ "lstrip": false,
814
+ "normalized": false,
815
+ "rstrip": false,
816
+ "single_word": false,
817
+ "special": true
818
+ },
819
+ "32099": {
820
+ "content": "<extra_id_0>",
821
+ "lstrip": false,
822
+ "normalized": false,
823
+ "rstrip": false,
824
+ "single_word": false,
825
+ "special": true
826
+ }
827
+ },
828
+ "additional_special_tokens": [
829
+ "<extra_id_0>",
830
+ "<extra_id_1>",
831
+ "<extra_id_2>",
832
+ "<extra_id_3>",
833
+ "<extra_id_4>",
834
+ "<extra_id_5>",
835
+ "<extra_id_6>",
836
+ "<extra_id_7>",
837
+ "<extra_id_8>",
838
+ "<extra_id_9>",
839
+ "<extra_id_10>",
840
+ "<extra_id_11>",
841
+ "<extra_id_12>",
842
+ "<extra_id_13>",
843
+ "<extra_id_14>",
844
+ "<extra_id_15>",
845
+ "<extra_id_16>",
846
+ "<extra_id_17>",
847
+ "<extra_id_18>",
848
+ "<extra_id_19>",
849
+ "<extra_id_20>",
850
+ "<extra_id_21>",
851
+ "<extra_id_22>",
852
+ "<extra_id_23>",
853
+ "<extra_id_24>",
854
+ "<extra_id_25>",
855
+ "<extra_id_26>",
856
+ "<extra_id_27>",
857
+ "<extra_id_28>",
858
+ "<extra_id_29>",
859
+ "<extra_id_30>",
860
+ "<extra_id_31>",
861
+ "<extra_id_32>",
862
+ "<extra_id_33>",
863
+ "<extra_id_34>",
864
+ "<extra_id_35>",
865
+ "<extra_id_36>",
866
+ "<extra_id_37>",
867
+ "<extra_id_38>",
868
+ "<extra_id_39>",
869
+ "<extra_id_40>",
870
+ "<extra_id_41>",
871
+ "<extra_id_42>",
872
+ "<extra_id_43>",
873
+ "<extra_id_44>",
874
+ "<extra_id_45>",
875
+ "<extra_id_46>",
876
+ "<extra_id_47>",
877
+ "<extra_id_48>",
878
+ "<extra_id_49>",
879
+ "<extra_id_50>",
880
+ "<extra_id_51>",
881
+ "<extra_id_52>",
882
+ "<extra_id_53>",
883
+ "<extra_id_54>",
884
+ "<extra_id_55>",
885
+ "<extra_id_56>",
886
+ "<extra_id_57>",
887
+ "<extra_id_58>",
888
+ "<extra_id_59>",
889
+ "<extra_id_60>",
890
+ "<extra_id_61>",
891
+ "<extra_id_62>",
892
+ "<extra_id_63>",
893
+ "<extra_id_64>",
894
+ "<extra_id_65>",
895
+ "<extra_id_66>",
896
+ "<extra_id_67>",
897
+ "<extra_id_68>",
898
+ "<extra_id_69>",
899
+ "<extra_id_70>",
900
+ "<extra_id_71>",
901
+ "<extra_id_72>",
902
+ "<extra_id_73>",
903
+ "<extra_id_74>",
904
+ "<extra_id_75>",
905
+ "<extra_id_76>",
906
+ "<extra_id_77>",
907
+ "<extra_id_78>",
908
+ "<extra_id_79>",
909
+ "<extra_id_80>",
910
+ "<extra_id_81>",
911
+ "<extra_id_82>",
912
+ "<extra_id_83>",
913
+ "<extra_id_84>",
914
+ "<extra_id_85>",
915
+ "<extra_id_86>",
916
+ "<extra_id_87>",
917
+ "<extra_id_88>",
918
+ "<extra_id_89>",
919
+ "<extra_id_90>",
920
+ "<extra_id_91>",
921
+ "<extra_id_92>",
922
+ "<extra_id_93>",
923
+ "<extra_id_94>",
924
+ "<extra_id_95>",
925
+ "<extra_id_96>",
926
+ "<extra_id_97>",
927
+ "<extra_id_98>",
928
+ "<extra_id_99>"
929
+ ],
930
+ "clean_up_tokenization_spaces": true,
931
+ "eos_token": "</s>",
932
+ "extra_ids": 100,
933
+ "model_max_length": 512,
934
+ "pad_token": "<pad>",
935
+ "sp_model_kwargs": {},
936
+ "tokenizer_class": "T5Tokenizer",
937
+ "unk_token": "<unk>"
938
+ }
train_results.json ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 3.83,
3
+ "train_loss": 0.16499798774719238,
4
+ "train_runtime": 3538.4305,
5
+ "train_samples": 16700,
6
+ "train_samples_per_second": 47.196,
7
+ "train_steps_per_second": 0.738
8
+ }
trainer_state.json ADDED
@@ -0,0 +1,530 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 0.24614077806472778,
3
+ "best_model_checkpoint": "../outputs/deductor-flant5-large/checkpoint-500",
4
+ "epoch": 3.8314176245210727,
5
+ "eval_steps": 50,
6
+ "global_step": 1000,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 0.1,
13
+ "learning_rate": 4.952107279693487e-05,
14
+ "loss": 0.5028,
15
+ "step": 25
16
+ },
17
+ {
18
+ "epoch": 0.19,
19
+ "learning_rate": 4.904214559386973e-05,
20
+ "loss": 0.306,
21
+ "step": 50
22
+ },
23
+ {
24
+ "epoch": 0.19,
25
+ "eval_gen_len": 11.22110552763819,
26
+ "eval_loss": 0.2959373891353607,
27
+ "eval_rouge1": 89.3028,
28
+ "eval_rouge2": 82.5127,
29
+ "eval_rougeL": 87.4173,
30
+ "eval_rougeLsum": 87.3544,
31
+ "eval_runtime": 58.3773,
32
+ "eval_samples_per_second": 40.906,
33
+ "eval_steps_per_second": 1.285,
34
+ "step": 50
35
+ },
36
+ {
37
+ "epoch": 0.29,
38
+ "learning_rate": 4.85632183908046e-05,
39
+ "loss": 0.2749,
40
+ "step": 75
41
+ },
42
+ {
43
+ "epoch": 0.38,
44
+ "learning_rate": 4.8084291187739464e-05,
45
+ "loss": 0.2774,
46
+ "step": 100
47
+ },
48
+ {
49
+ "epoch": 0.38,
50
+ "eval_gen_len": 11.257118927973199,
51
+ "eval_loss": 0.2716549038887024,
52
+ "eval_rouge1": 90.8414,
53
+ "eval_rouge2": 84.2378,
54
+ "eval_rougeL": 88.9385,
55
+ "eval_rougeLsum": 88.9058,
56
+ "eval_runtime": 58.4865,
57
+ "eval_samples_per_second": 40.83,
58
+ "eval_steps_per_second": 1.282,
59
+ "step": 100
60
+ },
61
+ {
62
+ "epoch": 0.48,
63
+ "learning_rate": 4.760536398467433e-05,
64
+ "loss": 0.2676,
65
+ "step": 125
66
+ },
67
+ {
68
+ "epoch": 0.57,
69
+ "learning_rate": 4.7126436781609195e-05,
70
+ "loss": 0.2366,
71
+ "step": 150
72
+ },
73
+ {
74
+ "epoch": 0.57,
75
+ "eval_gen_len": 11.208123953098827,
76
+ "eval_loss": 0.2612508535385132,
77
+ "eval_rouge1": 91.0152,
78
+ "eval_rouge2": 84.6687,
79
+ "eval_rougeL": 89.2107,
80
+ "eval_rougeLsum": 89.1735,
81
+ "eval_runtime": 58.169,
82
+ "eval_samples_per_second": 41.053,
83
+ "eval_steps_per_second": 1.289,
84
+ "step": 150
85
+ },
86
+ {
87
+ "epoch": 0.67,
88
+ "learning_rate": 4.6647509578544064e-05,
89
+ "loss": 0.274,
90
+ "step": 175
91
+ },
92
+ {
93
+ "epoch": 0.77,
94
+ "learning_rate": 4.616858237547893e-05,
95
+ "loss": 0.2166,
96
+ "step": 200
97
+ },
98
+ {
99
+ "epoch": 0.77,
100
+ "eval_gen_len": 11.280150753768844,
101
+ "eval_loss": 0.25851312279701233,
102
+ "eval_rouge1": 91.5215,
103
+ "eval_rouge2": 85.4308,
104
+ "eval_rougeL": 89.7742,
105
+ "eval_rougeLsum": 89.7422,
106
+ "eval_runtime": 58.2807,
107
+ "eval_samples_per_second": 40.974,
108
+ "eval_steps_per_second": 1.287,
109
+ "step": 200
110
+ },
111
+ {
112
+ "epoch": 0.86,
113
+ "learning_rate": 4.5689655172413794e-05,
114
+ "loss": 0.2063,
115
+ "step": 225
116
+ },
117
+ {
118
+ "epoch": 0.96,
119
+ "learning_rate": 4.5210727969348656e-05,
120
+ "loss": 0.22,
121
+ "step": 250
122
+ },
123
+ {
124
+ "epoch": 0.96,
125
+ "eval_gen_len": 11.265494137353434,
126
+ "eval_loss": 0.25169575214385986,
127
+ "eval_rouge1": 91.5587,
128
+ "eval_rouge2": 85.6107,
129
+ "eval_rougeL": 89.8835,
130
+ "eval_rougeLsum": 89.8621,
131
+ "eval_runtime": 58.1997,
132
+ "eval_samples_per_second": 41.031,
133
+ "eval_steps_per_second": 1.289,
134
+ "step": 250
135
+ },
136
+ {
137
+ "epoch": 1.05,
138
+ "learning_rate": 4.4731800766283525e-05,
139
+ "loss": 0.196,
140
+ "step": 275
141
+ },
142
+ {
143
+ "epoch": 1.15,
144
+ "learning_rate": 4.4252873563218394e-05,
145
+ "loss": 0.1564,
146
+ "step": 300
147
+ },
148
+ {
149
+ "epoch": 1.15,
150
+ "eval_gen_len": 11.20393634840871,
151
+ "eval_loss": 0.26295191049575806,
152
+ "eval_rouge1": 91.999,
153
+ "eval_rouge2": 86.0835,
154
+ "eval_rougeL": 90.3611,
155
+ "eval_rougeLsum": 90.3168,
156
+ "eval_runtime": 58.4367,
157
+ "eval_samples_per_second": 40.865,
158
+ "eval_steps_per_second": 1.283,
159
+ "step": 300
160
+ },
161
+ {
162
+ "epoch": 1.25,
163
+ "learning_rate": 4.3773946360153256e-05,
164
+ "loss": 0.1545,
165
+ "step": 325
166
+ },
167
+ {
168
+ "epoch": 1.34,
169
+ "learning_rate": 4.3295019157088125e-05,
170
+ "loss": 0.1803,
171
+ "step": 350
172
+ },
173
+ {
174
+ "epoch": 1.34,
175
+ "eval_gen_len": 11.246231155778894,
176
+ "eval_loss": 0.2546021044254303,
177
+ "eval_rouge1": 91.5183,
178
+ "eval_rouge2": 85.6214,
179
+ "eval_rougeL": 89.9752,
180
+ "eval_rougeLsum": 89.9323,
181
+ "eval_runtime": 58.1825,
182
+ "eval_samples_per_second": 41.043,
183
+ "eval_steps_per_second": 1.289,
184
+ "step": 350
185
+ },
186
+ {
187
+ "epoch": 1.44,
188
+ "learning_rate": 4.2816091954022994e-05,
189
+ "loss": 0.1793,
190
+ "step": 375
191
+ },
192
+ {
193
+ "epoch": 1.53,
194
+ "learning_rate": 4.2337164750957856e-05,
195
+ "loss": 0.1737,
196
+ "step": 400
197
+ },
198
+ {
199
+ "epoch": 1.53,
200
+ "eval_gen_len": 11.194304857621441,
201
+ "eval_loss": 0.24834655225276947,
202
+ "eval_rouge1": 91.8342,
203
+ "eval_rouge2": 86.0171,
204
+ "eval_rougeL": 90.3042,
205
+ "eval_rougeLsum": 90.2641,
206
+ "eval_runtime": 58.3123,
207
+ "eval_samples_per_second": 40.952,
208
+ "eval_steps_per_second": 1.286,
209
+ "step": 400
210
+ },
211
+ {
212
+ "epoch": 1.63,
213
+ "learning_rate": 4.185823754789272e-05,
214
+ "loss": 0.166,
215
+ "step": 425
216
+ },
217
+ {
218
+ "epoch": 1.72,
219
+ "learning_rate": 4.1379310344827587e-05,
220
+ "loss": 0.157,
221
+ "step": 450
222
+ },
223
+ {
224
+ "epoch": 1.72,
225
+ "eval_gen_len": 10.937604690117253,
226
+ "eval_loss": 0.24926304817199707,
227
+ "eval_rouge1": 91.6585,
228
+ "eval_rouge2": 85.4651,
229
+ "eval_rougeL": 90.0181,
230
+ "eval_rougeLsum": 89.9991,
231
+ "eval_runtime": 57.6625,
232
+ "eval_samples_per_second": 41.413,
233
+ "eval_steps_per_second": 1.301,
234
+ "step": 450
235
+ },
236
+ {
237
+ "epoch": 1.82,
238
+ "learning_rate": 4.0900383141762455e-05,
239
+ "loss": 0.1612,
240
+ "step": 475
241
+ },
242
+ {
243
+ "epoch": 1.92,
244
+ "learning_rate": 4.0421455938697324e-05,
245
+ "loss": 0.1561,
246
+ "step": 500
247
+ },
248
+ {
249
+ "epoch": 1.92,
250
+ "eval_gen_len": 11.20142378559464,
251
+ "eval_loss": 0.24614077806472778,
252
+ "eval_rouge1": 92.1213,
253
+ "eval_rouge2": 86.4281,
254
+ "eval_rougeL": 90.5846,
255
+ "eval_rougeLsum": 90.5294,
256
+ "eval_runtime": 58.4684,
257
+ "eval_samples_per_second": 40.843,
258
+ "eval_steps_per_second": 1.283,
259
+ "step": 500
260
+ },
261
+ {
262
+ "epoch": 2.01,
263
+ "learning_rate": 3.9942528735632186e-05,
264
+ "loss": 0.1472,
265
+ "step": 525
266
+ },
267
+ {
268
+ "epoch": 2.11,
269
+ "learning_rate": 3.9463601532567055e-05,
270
+ "loss": 0.1191,
271
+ "step": 550
272
+ },
273
+ {
274
+ "epoch": 2.11,
275
+ "eval_gen_len": 11.241624790619765,
276
+ "eval_loss": 0.2584824860095978,
277
+ "eval_rouge1": 92.4493,
278
+ "eval_rouge2": 86.6961,
279
+ "eval_rougeL": 90.9293,
280
+ "eval_rougeLsum": 90.8761,
281
+ "eval_runtime": 58.3545,
282
+ "eval_samples_per_second": 40.922,
283
+ "eval_steps_per_second": 1.285,
284
+ "step": 550
285
+ },
286
+ {
287
+ "epoch": 2.2,
288
+ "learning_rate": 3.898467432950192e-05,
289
+ "loss": 0.1252,
290
+ "step": 575
291
+ },
292
+ {
293
+ "epoch": 2.3,
294
+ "learning_rate": 3.850574712643678e-05,
295
+ "loss": 0.1134,
296
+ "step": 600
297
+ },
298
+ {
299
+ "epoch": 2.3,
300
+ "eval_gen_len": 11.16750418760469,
301
+ "eval_loss": 0.2633197009563446,
302
+ "eval_rouge1": 92.4707,
303
+ "eval_rouge2": 86.833,
304
+ "eval_rougeL": 90.9516,
305
+ "eval_rougeLsum": 90.9195,
306
+ "eval_runtime": 58.1972,
307
+ "eval_samples_per_second": 41.033,
308
+ "eval_steps_per_second": 1.289,
309
+ "step": 600
310
+ },
311
+ {
312
+ "epoch": 2.39,
313
+ "learning_rate": 3.802681992337165e-05,
314
+ "loss": 0.1128,
315
+ "step": 625
316
+ },
317
+ {
318
+ "epoch": 2.49,
319
+ "learning_rate": 3.7547892720306517e-05,
320
+ "loss": 0.1227,
321
+ "step": 650
322
+ },
323
+ {
324
+ "epoch": 2.49,
325
+ "eval_gen_len": 11.264237855946398,
326
+ "eval_loss": 0.25923022627830505,
327
+ "eval_rouge1": 92.2738,
328
+ "eval_rouge2": 86.5064,
329
+ "eval_rougeL": 90.7556,
330
+ "eval_rougeLsum": 90.6998,
331
+ "eval_runtime": 58.2266,
332
+ "eval_samples_per_second": 41.012,
333
+ "eval_steps_per_second": 1.288,
334
+ "step": 650
335
+ },
336
+ {
337
+ "epoch": 2.59,
338
+ "learning_rate": 3.7068965517241385e-05,
339
+ "loss": 0.1232,
340
+ "step": 675
341
+ },
342
+ {
343
+ "epoch": 2.68,
344
+ "learning_rate": 3.659003831417625e-05,
345
+ "loss": 0.1175,
346
+ "step": 700
347
+ },
348
+ {
349
+ "epoch": 2.68,
350
+ "eval_gen_len": 11.17001675041876,
351
+ "eval_loss": 0.2656923830509186,
352
+ "eval_rouge1": 92.0861,
353
+ "eval_rouge2": 86.2203,
354
+ "eval_rougeL": 90.6168,
355
+ "eval_rougeLsum": 90.5657,
356
+ "eval_runtime": 58.0131,
357
+ "eval_samples_per_second": 41.163,
358
+ "eval_steps_per_second": 1.293,
359
+ "step": 700
360
+ },
361
+ {
362
+ "epoch": 2.78,
363
+ "learning_rate": 3.611111111111111e-05,
364
+ "loss": 0.1095,
365
+ "step": 725
366
+ },
367
+ {
368
+ "epoch": 2.87,
369
+ "learning_rate": 3.563218390804598e-05,
370
+ "loss": 0.1132,
371
+ "step": 750
372
+ },
373
+ {
374
+ "epoch": 2.87,
375
+ "eval_gen_len": 11.212311557788945,
376
+ "eval_loss": 0.26437509059906006,
377
+ "eval_rouge1": 92.3834,
378
+ "eval_rouge2": 86.7237,
379
+ "eval_rougeL": 90.8761,
380
+ "eval_rougeLsum": 90.8389,
381
+ "eval_runtime": 58.1768,
382
+ "eval_samples_per_second": 41.047,
383
+ "eval_steps_per_second": 1.289,
384
+ "step": 750
385
+ },
386
+ {
387
+ "epoch": 2.97,
388
+ "learning_rate": 3.515325670498085e-05,
389
+ "loss": 0.1124,
390
+ "step": 775
391
+ },
392
+ {
393
+ "epoch": 3.07,
394
+ "learning_rate": 3.467432950191571e-05,
395
+ "loss": 0.1097,
396
+ "step": 800
397
+ },
398
+ {
399
+ "epoch": 3.07,
400
+ "eval_gen_len": 11.1821608040201,
401
+ "eval_loss": 0.2691878378391266,
402
+ "eval_rouge1": 92.3356,
403
+ "eval_rouge2": 86.7021,
404
+ "eval_rougeL": 90.8717,
405
+ "eval_rougeLsum": 90.8185,
406
+ "eval_runtime": 58.2708,
407
+ "eval_samples_per_second": 40.981,
408
+ "eval_steps_per_second": 1.287,
409
+ "step": 800
410
+ },
411
+ {
412
+ "epoch": 3.16,
413
+ "learning_rate": 3.419540229885058e-05,
414
+ "loss": 0.0874,
415
+ "step": 825
416
+ },
417
+ {
418
+ "epoch": 3.26,
419
+ "learning_rate": 3.371647509578545e-05,
420
+ "loss": 0.0949,
421
+ "step": 850
422
+ },
423
+ {
424
+ "epoch": 3.26,
425
+ "eval_gen_len": 11.278475711892797,
426
+ "eval_loss": 0.26897767186164856,
427
+ "eval_rouge1": 92.5746,
428
+ "eval_rouge2": 87.001,
429
+ "eval_rougeL": 91.1734,
430
+ "eval_rougeLsum": 91.1222,
431
+ "eval_runtime": 58.645,
432
+ "eval_samples_per_second": 40.72,
433
+ "eval_steps_per_second": 1.279,
434
+ "step": 850
435
+ },
436
+ {
437
+ "epoch": 3.35,
438
+ "learning_rate": 3.323754789272031e-05,
439
+ "loss": 0.0901,
440
+ "step": 875
441
+ },
442
+ {
443
+ "epoch": 3.45,
444
+ "learning_rate": 3.275862068965517e-05,
445
+ "loss": 0.0813,
446
+ "step": 900
447
+ },
448
+ {
449
+ "epoch": 3.45,
450
+ "eval_gen_len": 11.22571189279732,
451
+ "eval_loss": 0.2874927222728729,
452
+ "eval_rouge1": 92.5641,
453
+ "eval_rouge2": 86.9813,
454
+ "eval_rougeL": 91.0881,
455
+ "eval_rougeLsum": 91.0411,
456
+ "eval_runtime": 58.3729,
457
+ "eval_samples_per_second": 40.909,
458
+ "eval_steps_per_second": 1.285,
459
+ "step": 900
460
+ },
461
+ {
462
+ "epoch": 3.54,
463
+ "learning_rate": 3.227969348659004e-05,
464
+ "loss": 0.1005,
465
+ "step": 925
466
+ },
467
+ {
468
+ "epoch": 3.64,
469
+ "learning_rate": 3.180076628352491e-05,
470
+ "loss": 0.0861,
471
+ "step": 950
472
+ },
473
+ {
474
+ "epoch": 3.64,
475
+ "eval_gen_len": 11.21356783919598,
476
+ "eval_loss": 0.280032217502594,
477
+ "eval_rouge1": 92.4738,
478
+ "eval_rouge2": 86.9379,
479
+ "eval_rougeL": 91.0384,
480
+ "eval_rougeLsum": 90.9995,
481
+ "eval_runtime": 58.3261,
482
+ "eval_samples_per_second": 40.942,
483
+ "eval_steps_per_second": 1.286,
484
+ "step": 950
485
+ },
486
+ {
487
+ "epoch": 3.74,
488
+ "learning_rate": 3.132183908045977e-05,
489
+ "loss": 0.0828,
490
+ "step": 975
491
+ },
492
+ {
493
+ "epoch": 3.83,
494
+ "learning_rate": 3.084291187739464e-05,
495
+ "loss": 0.0879,
496
+ "step": 1000
497
+ },
498
+ {
499
+ "epoch": 3.83,
500
+ "eval_gen_len": 11.23031825795645,
501
+ "eval_loss": 0.27702075242996216,
502
+ "eval_rouge1": 92.6025,
503
+ "eval_rouge2": 87.105,
504
+ "eval_rougeL": 91.1632,
505
+ "eval_rougeLsum": 91.1292,
506
+ "eval_runtime": 58.0228,
507
+ "eval_samples_per_second": 41.156,
508
+ "eval_steps_per_second": 1.293,
509
+ "step": 1000
510
+ },
511
+ {
512
+ "epoch": 3.83,
513
+ "step": 1000,
514
+ "total_flos": 2.492186999051059e+16,
515
+ "train_loss": 0.16499798774719238,
516
+ "train_runtime": 3538.4305,
517
+ "train_samples_per_second": 47.196,
518
+ "train_steps_per_second": 0.738
519
+ }
520
+ ],
521
+ "logging_steps": 25,
522
+ "max_steps": 2610,
523
+ "num_input_tokens_seen": 0,
524
+ "num_train_epochs": 10,
525
+ "save_steps": 50,
526
+ "total_flos": 2.492186999051059e+16,
527
+ "train_batch_size": 16,
528
+ "trial_name": null,
529
+ "trial_params": null
530
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:59a446301691724c7427d3b42fb36d0aa39b542460c8c234af0d4c8eabfe4b41
3
+ size 5883