Training in progress, epoch 0
Browse files- config.json +8 -8
- pytorch_model.bin +1 -1
config.json
CHANGED
@@ -6,31 +6,31 @@
|
|
6 |
],
|
7 |
"decoder": {
|
8 |
"_name_or_path": "",
|
9 |
-
"activation_dropout": 0.
|
10 |
"activation_function": "gelu",
|
11 |
"add_cross_attention": true,
|
12 |
"add_final_layer_norm": true,
|
13 |
"architectures": null,
|
14 |
-
"attention_dropout": 0.
|
15 |
"bad_words_ids": null,
|
16 |
"begin_suppress_tokens": null,
|
17 |
"bos_token_id": 0,
|
18 |
"chunk_size_feed_forward": 0,
|
19 |
-
"classifier_dropout": 0.
|
20 |
"cross_attention_hidden_size": null,
|
21 |
"d_model": 1024,
|
22 |
"decoder_attention_heads": 16,
|
23 |
"decoder_ffn_dim": 4096,
|
24 |
-
"decoder_layerdrop": 0.
|
25 |
"decoder_layers": 4,
|
26 |
"decoder_start_token_id": null,
|
27 |
"diversity_penalty": 0.0,
|
28 |
"do_sample": false,
|
29 |
-
"dropout": 0.
|
30 |
"early_stopping": false,
|
31 |
"encoder_attention_heads": 16,
|
32 |
"encoder_ffn_dim": 4096,
|
33 |
-
"encoder_layerdrop": 0.
|
34 |
"encoder_layers": 12,
|
35 |
"encoder_no_repeat_ngram_size": 0,
|
36 |
"eos_token_id": 2,
|
@@ -94,7 +94,7 @@
|
|
94 |
"_name_or_path": "",
|
95 |
"add_cross_attention": false,
|
96 |
"architectures": null,
|
97 |
-
"attention_probs_dropout_prob": 0.
|
98 |
"bad_words_ids": null,
|
99 |
"begin_suppress_tokens": null,
|
100 |
"bos_token_id": null,
|
@@ -119,7 +119,7 @@
|
|
119 |
"forced_bos_token_id": null,
|
120 |
"forced_eos_token_id": null,
|
121 |
"hidden_act": "gelu",
|
122 |
-
"hidden_dropout_prob": 0.
|
123 |
"hidden_size": 1024,
|
124 |
"id2label": {
|
125 |
"0": "LABEL_0",
|
|
|
6 |
],
|
7 |
"decoder": {
|
8 |
"_name_or_path": "",
|
9 |
+
"activation_dropout": 0.0,
|
10 |
"activation_function": "gelu",
|
11 |
"add_cross_attention": true,
|
12 |
"add_final_layer_norm": true,
|
13 |
"architectures": null,
|
14 |
+
"attention_dropout": 0.0,
|
15 |
"bad_words_ids": null,
|
16 |
"begin_suppress_tokens": null,
|
17 |
"bos_token_id": 0,
|
18 |
"chunk_size_feed_forward": 0,
|
19 |
+
"classifier_dropout": 0.0,
|
20 |
"cross_attention_hidden_size": null,
|
21 |
"d_model": 1024,
|
22 |
"decoder_attention_heads": 16,
|
23 |
"decoder_ffn_dim": 4096,
|
24 |
+
"decoder_layerdrop": 0.0,
|
25 |
"decoder_layers": 4,
|
26 |
"decoder_start_token_id": null,
|
27 |
"diversity_penalty": 0.0,
|
28 |
"do_sample": false,
|
29 |
+
"dropout": 0.1,
|
30 |
"early_stopping": false,
|
31 |
"encoder_attention_heads": 16,
|
32 |
"encoder_ffn_dim": 4096,
|
33 |
+
"encoder_layerdrop": 0.0,
|
34 |
"encoder_layers": 12,
|
35 |
"encoder_no_repeat_ngram_size": 0,
|
36 |
"eos_token_id": 2,
|
|
|
94 |
"_name_or_path": "",
|
95 |
"add_cross_attention": false,
|
96 |
"architectures": null,
|
97 |
+
"attention_probs_dropout_prob": 0.0,
|
98 |
"bad_words_ids": null,
|
99 |
"begin_suppress_tokens": null,
|
100 |
"bos_token_id": null,
|
|
|
119 |
"forced_bos_token_id": null,
|
120 |
"forced_eos_token_id": null,
|
121 |
"hidden_act": "gelu",
|
122 |
+
"hidden_dropout_prob": 0.0,
|
123 |
"hidden_size": 1024,
|
124 |
"id2label": {
|
125 |
"0": "LABEL_0",
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 806283643
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dccf107f3c24acbb028775dbeaa3726da6846fa60b7bd9e0684fadf748eba8a3
|
3 |
size 806283643
|