thisisiron commited on
Commit
34ef755
·
1 Parent(s): b12dc77

add korclip

Browse files
Files changed (7) hide show
  1. config.json +168 -0
  2. log.txt +90 -0
  3. model.safetensors +3 -0
  4. special_tokens_map.json +51 -0
  5. tokenizer.json +0 -0
  6. tokenizer_config.json +59 -0
  7. vocab.txt +0 -0
config.json ADDED
@@ -0,0 +1,168 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "VisionTextDualEncoderModel"
4
+ ],
5
+ "logit_scale_init_value": 2.6592,
6
+ "model_type": "vision-text-dual-encoder",
7
+ "projection_dim": 512,
8
+ "text_config": {
9
+ "_name_or_path": "klue/roberta-large",
10
+ "add_cross_attention": false,
11
+ "architectures": [
12
+ "RobertaForMaskedLM"
13
+ ],
14
+ "attention_probs_dropout_prob": 0.1,
15
+ "bad_words_ids": null,
16
+ "begin_suppress_tokens": null,
17
+ "bos_token_id": 0,
18
+ "chunk_size_feed_forward": 0,
19
+ "classifier_dropout": null,
20
+ "cross_attention_hidden_size": null,
21
+ "decoder_start_token_id": null,
22
+ "diversity_penalty": 0.0,
23
+ "do_sample": false,
24
+ "early_stopping": false,
25
+ "encoder_no_repeat_ngram_size": 0,
26
+ "eos_token_id": 2,
27
+ "exponential_decay_length_penalty": null,
28
+ "finetuning_task": null,
29
+ "forced_bos_token_id": null,
30
+ "forced_eos_token_id": null,
31
+ "gradient_checkpointing": false,
32
+ "hidden_act": "gelu",
33
+ "hidden_dropout_prob": 0.1,
34
+ "hidden_size": 1024,
35
+ "id2label": {
36
+ "0": "LABEL_0",
37
+ "1": "LABEL_1"
38
+ },
39
+ "initializer_range": 0.02,
40
+ "intermediate_size": 4096,
41
+ "is_decoder": false,
42
+ "is_encoder_decoder": false,
43
+ "label2id": {
44
+ "LABEL_0": 0,
45
+ "LABEL_1": 1
46
+ },
47
+ "layer_norm_eps": 1e-05,
48
+ "length_penalty": 1.0,
49
+ "max_length": 20,
50
+ "max_position_embeddings": 514,
51
+ "min_length": 0,
52
+ "model_type": "roberta",
53
+ "no_repeat_ngram_size": 0,
54
+ "num_attention_heads": 16,
55
+ "num_beam_groups": 1,
56
+ "num_beams": 1,
57
+ "num_hidden_layers": 24,
58
+ "num_return_sequences": 1,
59
+ "output_attentions": false,
60
+ "output_hidden_states": false,
61
+ "output_scores": false,
62
+ "pad_token_id": 1,
63
+ "position_embedding_type": "absolute",
64
+ "prefix": null,
65
+ "problem_type": null,
66
+ "pruned_heads": {},
67
+ "remove_invalid_values": false,
68
+ "repetition_penalty": 1.0,
69
+ "return_dict": true,
70
+ "return_dict_in_generate": false,
71
+ "sep_token_id": null,
72
+ "suppress_tokens": null,
73
+ "task_specific_params": null,
74
+ "temperature": 1.0,
75
+ "tf_legacy_loss": false,
76
+ "tie_encoder_decoder": false,
77
+ "tie_word_embeddings": true,
78
+ "tokenizer_class": "BertTokenizer",
79
+ "top_k": 50,
80
+ "top_p": 1.0,
81
+ "torch_dtype": null,
82
+ "torchscript": false,
83
+ "type_vocab_size": 1,
84
+ "typical_p": 1.0,
85
+ "use_bfloat16": false,
86
+ "use_cache": true,
87
+ "vocab_size": 32000
88
+ },
89
+ "torch_dtype": "float32",
90
+ "transformers_version": "4.45.0",
91
+ "vision_config": {
92
+ "_name_or_path": "openai/clip-vit-base-patch32",
93
+ "add_cross_attention": false,
94
+ "architectures": null,
95
+ "attention_dropout": 0.0,
96
+ "bad_words_ids": null,
97
+ "begin_suppress_tokens": null,
98
+ "bos_token_id": null,
99
+ "chunk_size_feed_forward": 0,
100
+ "cross_attention_hidden_size": null,
101
+ "decoder_start_token_id": null,
102
+ "diversity_penalty": 0.0,
103
+ "do_sample": false,
104
+ "dropout": 0.0,
105
+ "early_stopping": false,
106
+ "encoder_no_repeat_ngram_size": 0,
107
+ "eos_token_id": null,
108
+ "exponential_decay_length_penalty": null,
109
+ "finetuning_task": null,
110
+ "forced_bos_token_id": null,
111
+ "forced_eos_token_id": null,
112
+ "hidden_act": "quick_gelu",
113
+ "hidden_size": 768,
114
+ "id2label": {
115
+ "0": "LABEL_0",
116
+ "1": "LABEL_1"
117
+ },
118
+ "image_size": 224,
119
+ "initializer_factor": 1.0,
120
+ "initializer_range": 0.02,
121
+ "intermediate_size": 3072,
122
+ "is_decoder": false,
123
+ "is_encoder_decoder": false,
124
+ "label2id": {
125
+ "LABEL_0": 0,
126
+ "LABEL_1": 1
127
+ },
128
+ "layer_norm_eps": 1e-05,
129
+ "length_penalty": 1.0,
130
+ "max_length": 20,
131
+ "min_length": 0,
132
+ "model_type": "clip_vision_model",
133
+ "no_repeat_ngram_size": 0,
134
+ "num_attention_heads": 12,
135
+ "num_beam_groups": 1,
136
+ "num_beams": 1,
137
+ "num_channels": 3,
138
+ "num_hidden_layers": 12,
139
+ "num_return_sequences": 1,
140
+ "output_attentions": false,
141
+ "output_hidden_states": false,
142
+ "output_scores": false,
143
+ "pad_token_id": null,
144
+ "patch_size": 32,
145
+ "prefix": null,
146
+ "problem_type": null,
147
+ "projection_dim": 512,
148
+ "pruned_heads": {},
149
+ "remove_invalid_values": false,
150
+ "repetition_penalty": 1.0,
151
+ "return_dict": true,
152
+ "return_dict_in_generate": false,
153
+ "sep_token_id": null,
154
+ "suppress_tokens": null,
155
+ "task_specific_params": null,
156
+ "temperature": 1.0,
157
+ "tf_legacy_loss": false,
158
+ "tie_encoder_decoder": false,
159
+ "tie_word_embeddings": true,
160
+ "tokenizer_class": null,
161
+ "top_k": 50,
162
+ "top_p": 1.0,
163
+ "torch_dtype": null,
164
+ "torchscript": false,
165
+ "typical_p": 1.0,
166
+ "use_bfloat16": false
167
+ }
168
+ }
log.txt ADDED
@@ -0,0 +1,90 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ step: Resize image (224)
2
+ ***** Running training *****
3
+ Num examples = 414112
4
+ Num Epochs = 40
5
+ Instantaneous batch size per device = 240
6
+ Total train batch size = 240
7
+ Total optimization steps = 69000
8
+ Epoch 1/40 - Train loss: 1.1374
9
+ Epoch 1/40 - Top 1: 61.3591, Top 5: 95.0198
10
+ Model saved to output/2024-10-27T04-15-51/bset
11
+ Epoch 2/40 - Train loss: 0.6764
12
+ Epoch 2/40 - Top 1: 57.1230, Top 5: 94.3750
13
+ Epoch 3/40 - Train loss: 0.5364
14
+ Epoch 3/40 - Top 1: 62.6587, Top 5: 95.5655
15
+ Model saved to output/2024-10-27T04-15-51/bset
16
+ Epoch 4/40 - Train loss: 0.4502
17
+ Epoch 4/40 - Top 1: 59.2560, Top 5: 93.7202
18
+ Epoch 5/40 - Train loss: 0.3864
19
+ Epoch 5/40 - Top 1: 62.6389, Top 5: 94.5933
20
+ Epoch 6/40 - Train loss: 0.3365
21
+ Epoch 6/40 - Top 1: 54.2956, Top 5: 92.0734
22
+ Epoch 7/40 - Train loss: 0.2953
23
+ Epoch 7/40 - Top 1: 55.5952, Top 5: 91.4087
24
+ Epoch 8/40 - Train loss: 0.2609
25
+ Epoch 8/40 - Top 1: 59.7222, Top 5: 92.5694
26
+ Epoch 9/40 - Train loss: 0.2329
27
+ Epoch 9/40 - Top 1: 56.4583, Top 5: 93.7500
28
+ Epoch 10/40 - Train loss: 0.2086
29
+ Epoch 10/40 - Top 1: 56.5873, Top 5: 91.5575
30
+ Epoch 11/40 - Train loss: 0.1881
31
+ Epoch 11/40 - Top 1: 60.6944, Top 5: 92.2421
32
+ Epoch 12/40 - Train loss: 0.1713
33
+ Epoch 12/40 - Top 1: 49.0774, Top 5: 87.1627
34
+ Epoch 13/40 - Train loss: 0.1557
35
+ Epoch 13/40 - Top 1: 54.2758, Top 5: 89.9107
36
+ Epoch 14/40 - Train loss: 0.1435
37
+ Epoch 14/40 - Top 1: 53.4524, Top 5: 90.4067
38
+ Epoch 15/40 - Train loss: 0.1307
39
+ Epoch 15/40 - Top 1: 49.0079, Top 5: 87.0437
40
+ Epoch 16/40 - Train loss: 0.1176
41
+ Epoch 16/40 - Top 1: 55.2183, Top 5: 88.3135
42
+ Epoch 17/40 - Train loss: 0.1087
43
+ Epoch 17/40 - Top 1: 57.4802, Top 5: 90.7440
44
+ Epoch 18/40 - Train loss: 0.1013
45
+ Epoch 18/40 - Top 1: 53.0258, Top 5: 89.9107
46
+ Epoch 19/40 - Train loss: 0.0918
47
+ Epoch 19/40 - Top 1: 53.1647, Top 5: 90.4067
48
+ Epoch 20/40 - Train loss: 0.0857
49
+ Epoch 20/40 - Top 1: 50.6250, Top 5: 89.8313
50
+ Epoch 21/40 - Train loss: 0.0792
51
+ Epoch 21/40 - Top 1: 46.8750, Top 5: 86.8254
52
+ Epoch 22/40 - Train loss: 0.0721
53
+ Epoch 22/40 - Top 1: 45.2976, Top 5: 87.7778
54
+ Epoch 23/40 - Train loss: 0.0669
55
+ Epoch 23/40 - Top 1: 51.2202, Top 5: 89.5040
56
+ Epoch 24/40 - Train loss: 0.0623
57
+ Epoch 24/40 - Top 1: 47.9960, Top 5: 86.5675
58
+ Epoch 25/40 - Train loss: 0.0578
59
+ Epoch 25/40 - Top 1: 44.4345, Top 5: 87.3810
60
+ Epoch 26/40 - Train loss: 0.0534
61
+ Epoch 26/40 - Top 1: 49.0179, Top 5: 88.8294
62
+ Epoch 27/40 - Train loss: 0.0503
63
+ Epoch 27/40 - Top 1: 51.5476, Top 5: 89.5040
64
+ Epoch 28/40 - Train loss: 0.0460
65
+ Epoch 28/40 - Top 1: 55.8333, Top 5: 89.9107
66
+ Epoch 29/40 - Train loss: 0.0423
67
+ Epoch 29/40 - Top 1: 54.1270, Top 5: 89.8313
68
+ Epoch 30/40 - Train loss: 0.0392
69
+ Epoch 30/40 - Top 1: 50.6647, Top 5: 88.8492
70
+ Epoch 31/40 - Train loss: 0.0375
71
+ Epoch 31/40 - Top 1: 53.1448, Top 5: 89.0873
72
+ Epoch 32/40 - Train loss: 0.0336
73
+ Epoch 32/40 - Top 1: 50.9921, Top 5: 88.8889
74
+ Epoch 33/40 - Train loss: 0.0316
75
+ Epoch 33/40 - Top 1: 45.9325, Top 5: 84.7321
76
+ Epoch 34/40 - Train loss: 0.0289
77
+ Epoch 34/40 - Top 1: 50.4762, Top 5: 88.7004
78
+ Epoch 35/40 - Train loss: 0.0273
79
+ Epoch 35/40 - Top 1: 48.1250, Top 5: 87.3710
80
+ Epoch 36/40 - Train loss: 0.0248
81
+ Epoch 36/40 - Top 1: 45.8829, Top 5: 86.1111
82
+ Epoch 37/40 - Train loss: 0.0232
83
+ Epoch 37/40 - Top 1: 45.9425, Top 5: 85.1488
84
+ Epoch 38/40 - Train loss: 0.0216
85
+ Epoch 38/40 - Top 1: 46.0020, Top 5: 86.2698
86
+ Epoch 39/40 - Train loss: 0.0200
87
+ Epoch 39/40 - Top 1: 47.0139, Top 5: 86.3095
88
+ Epoch 40/40 - Train loss: 0.0193
89
+ Epoch 40/40 - Top 1: 47.5694, Top 5: 86.4286
90
+ Model saved to output/2024-10-27T04-15-51/last
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a5f021d992e7dbfa9f9388ba7414b80ff0e440fe7ba3c2b0a3b6cbad6d3e4086
3
+ size 1700195908
special_tokens_map.json ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "cls_token": {
10
+ "content": "[CLS]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "eos_token": {
17
+ "content": "[SEP]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "mask_token": {
24
+ "content": "[MASK]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "pad_token": {
31
+ "content": "[PAD]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ },
37
+ "sep_token": {
38
+ "content": "[SEP]",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false
43
+ },
44
+ "unk_token": {
45
+ "content": "[UNK]",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false
50
+ }
51
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,59 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[CLS]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "[PAD]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "[SEP]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "3": {
28
+ "content": "[UNK]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "4": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
+ "bos_token": "[CLS]",
45
+ "clean_up_tokenization_spaces": true,
46
+ "cls_token": "[CLS]",
47
+ "do_basic_tokenize": true,
48
+ "do_lower_case": false,
49
+ "eos_token": "[SEP]",
50
+ "mask_token": "[MASK]",
51
+ "model_max_length": 512,
52
+ "never_split": null,
53
+ "pad_token": "[PAD]",
54
+ "sep_token": "[SEP]",
55
+ "strip_accents": null,
56
+ "tokenize_chinese_chars": true,
57
+ "tokenizer_class": "BertTokenizer",
58
+ "unk_token": "[UNK]"
59
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff