PyWebSol commited on
Commit
dd03d0c
·
verified ·
1 Parent(s): 0c28512

Upload DetrForObjectDetection

Browse files
Files changed (3) hide show
  1. README.md +199 -0
  2. config.json +303 -0
  3. model.safetensors +3 -0
README.md ADDED
@@ -0,0 +1,199 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ library_name: transformers
3
+ tags: []
4
+ ---
5
+
6
+ # Model Card for Model ID
7
+
8
+ <!-- Provide a quick summary of what the model is/does. -->
9
+
10
+
11
+
12
+ ## Model Details
13
+
14
+ ### Model Description
15
+
16
+ <!-- Provide a longer summary of what this model is. -->
17
+
18
+ This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated.
19
+
20
+ - **Developed by:** [More Information Needed]
21
+ - **Funded by [optional]:** [More Information Needed]
22
+ - **Shared by [optional]:** [More Information Needed]
23
+ - **Model type:** [More Information Needed]
24
+ - **Language(s) (NLP):** [More Information Needed]
25
+ - **License:** [More Information Needed]
26
+ - **Finetuned from model [optional]:** [More Information Needed]
27
+
28
+ ### Model Sources [optional]
29
+
30
+ <!-- Provide the basic links for the model. -->
31
+
32
+ - **Repository:** [More Information Needed]
33
+ - **Paper [optional]:** [More Information Needed]
34
+ - **Demo [optional]:** [More Information Needed]
35
+
36
+ ## Uses
37
+
38
+ <!-- Address questions around how the model is intended to be used, including the foreseeable users of the model and those affected by the model. -->
39
+
40
+ ### Direct Use
41
+
42
+ <!-- This section is for the model use without fine-tuning or plugging into a larger ecosystem/app. -->
43
+
44
+ [More Information Needed]
45
+
46
+ ### Downstream Use [optional]
47
+
48
+ <!-- This section is for the model use when fine-tuned for a task, or when plugged into a larger ecosystem/app -->
49
+
50
+ [More Information Needed]
51
+
52
+ ### Out-of-Scope Use
53
+
54
+ <!-- This section addresses misuse, malicious use, and uses that the model will not work well for. -->
55
+
56
+ [More Information Needed]
57
+
58
+ ## Bias, Risks, and Limitations
59
+
60
+ <!-- This section is meant to convey both technical and sociotechnical limitations. -->
61
+
62
+ [More Information Needed]
63
+
64
+ ### Recommendations
65
+
66
+ <!-- This section is meant to convey recommendations with respect to the bias, risk, and technical limitations. -->
67
+
68
+ Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
69
+
70
+ ## How to Get Started with the Model
71
+
72
+ Use the code below to get started with the model.
73
+
74
+ [More Information Needed]
75
+
76
+ ## Training Details
77
+
78
+ ### Training Data
79
+
80
+ <!-- This should link to a Dataset Card, perhaps with a short stub of information on what the training data is all about as well as documentation related to data pre-processing or additional filtering. -->
81
+
82
+ [More Information Needed]
83
+
84
+ ### Training Procedure
85
+
86
+ <!-- This relates heavily to the Technical Specifications. Content here should link to that section when it is relevant to the training procedure. -->
87
+
88
+ #### Preprocessing [optional]
89
+
90
+ [More Information Needed]
91
+
92
+
93
+ #### Training Hyperparameters
94
+
95
+ - **Training regime:** [More Information Needed] <!--fp32, fp16 mixed precision, bf16 mixed precision, bf16 non-mixed precision, fp16 non-mixed precision, fp8 mixed precision -->
96
+
97
+ #### Speeds, Sizes, Times [optional]
98
+
99
+ <!-- This section provides information about throughput, start/end time, checkpoint size if relevant, etc. -->
100
+
101
+ [More Information Needed]
102
+
103
+ ## Evaluation
104
+
105
+ <!-- This section describes the evaluation protocols and provides the results. -->
106
+
107
+ ### Testing Data, Factors & Metrics
108
+
109
+ #### Testing Data
110
+
111
+ <!-- This should link to a Dataset Card if possible. -->
112
+
113
+ [More Information Needed]
114
+
115
+ #### Factors
116
+
117
+ <!-- These are the things the evaluation is disaggregating by, e.g., subpopulations or domains. -->
118
+
119
+ [More Information Needed]
120
+
121
+ #### Metrics
122
+
123
+ <!-- These are the evaluation metrics being used, ideally with a description of why. -->
124
+
125
+ [More Information Needed]
126
+
127
+ ### Results
128
+
129
+ [More Information Needed]
130
+
131
+ #### Summary
132
+
133
+
134
+
135
+ ## Model Examination [optional]
136
+
137
+ <!-- Relevant interpretability work for the model goes here -->
138
+
139
+ [More Information Needed]
140
+
141
+ ## Environmental Impact
142
+
143
+ <!-- Total emissions (in grams of CO2eq) and additional considerations, such as electricity usage, go here. Edit the suggested text below accordingly -->
144
+
145
+ Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
146
+
147
+ - **Hardware Type:** [More Information Needed]
148
+ - **Hours used:** [More Information Needed]
149
+ - **Cloud Provider:** [More Information Needed]
150
+ - **Compute Region:** [More Information Needed]
151
+ - **Carbon Emitted:** [More Information Needed]
152
+
153
+ ## Technical Specifications [optional]
154
+
155
+ ### Model Architecture and Objective
156
+
157
+ [More Information Needed]
158
+
159
+ ### Compute Infrastructure
160
+
161
+ [More Information Needed]
162
+
163
+ #### Hardware
164
+
165
+ [More Information Needed]
166
+
167
+ #### Software
168
+
169
+ [More Information Needed]
170
+
171
+ ## Citation [optional]
172
+
173
+ <!-- If there is a paper or blog post introducing the model, the APA and Bibtex information for that should go in this section. -->
174
+
175
+ **BibTeX:**
176
+
177
+ [More Information Needed]
178
+
179
+ **APA:**
180
+
181
+ [More Information Needed]
182
+
183
+ ## Glossary [optional]
184
+
185
+ <!-- If relevant, include terms and calculations in this section that can help readers understand the model or model card. -->
186
+
187
+ [More Information Needed]
188
+
189
+ ## More Information [optional]
190
+
191
+ [More Information Needed]
192
+
193
+ ## Model Card Authors [optional]
194
+
195
+ [More Information Needed]
196
+
197
+ ## Model Card Contact
198
+
199
+ [More Information Needed]
config.json ADDED
@@ -0,0 +1,303 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "facebook/detr-resnet-50",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "relu",
5
+ "architectures": [
6
+ "DetrForObjectDetection"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "auxiliary_loss": false,
10
+ "backbone": "resnet50",
11
+ "backbone_config": null,
12
+ "backbone_kwargs": {
13
+ "in_chans": 3,
14
+ "out_indices": [
15
+ 1,
16
+ 2,
17
+ 3,
18
+ 4
19
+ ]
20
+ },
21
+ "bbox_cost": 5,
22
+ "bbox_loss_coefficient": 5,
23
+ "class_cost": 1,
24
+ "classifier_dropout": 0.0,
25
+ "d_model": 256,
26
+ "decoder_attention_heads": 8,
27
+ "decoder_ffn_dim": 2048,
28
+ "decoder_layerdrop": 0.0,
29
+ "decoder_layers": 6,
30
+ "dice_loss_coefficient": 1,
31
+ "dilation": false,
32
+ "dropout": 0.1,
33
+ "encoder_attention_heads": 8,
34
+ "encoder_ffn_dim": 2048,
35
+ "encoder_layerdrop": 0.0,
36
+ "encoder_layers": 6,
37
+ "eos_coefficient": 0.1,
38
+ "giou_cost": 2,
39
+ "giou_loss_coefficient": 2,
40
+ "id2label": {
41
+ "0": "[PAD]",
42
+ "1": "[UNK]",
43
+ "2": "[CLS]",
44
+ "3": "[SEP]",
45
+ "4": "[MASK]",
46
+ "5": " ",
47
+ "6": "A",
48
+ "7": "B",
49
+ "8": "C",
50
+ "9": "D",
51
+ "10": "E",
52
+ "11": "F",
53
+ "12": "G",
54
+ "13": "H",
55
+ "14": "I",
56
+ "15": "J",
57
+ "16": "K",
58
+ "17": "L",
59
+ "18": "M",
60
+ "19": "N",
61
+ "20": "O",
62
+ "21": "P",
63
+ "22": "Q",
64
+ "23": "R",
65
+ "24": "S",
66
+ "25": "T",
67
+ "26": "U",
68
+ "27": "V",
69
+ "28": "W",
70
+ "29": "X",
71
+ "30": "Y",
72
+ "31": "Z",
73
+ "32": "a",
74
+ "33": "b",
75
+ "34": "c",
76
+ "35": "d",
77
+ "36": "e",
78
+ "37": "f",
79
+ "38": "g",
80
+ "39": "h",
81
+ "40": "i",
82
+ "41": "j",
83
+ "42": "k",
84
+ "43": "l",
85
+ "44": "m",
86
+ "45": "n",
87
+ "46": "o",
88
+ "47": "p",
89
+ "48": "q",
90
+ "49": "r",
91
+ "50": "s",
92
+ "51": "t",
93
+ "52": "u",
94
+ "53": "v",
95
+ "54": "w",
96
+ "55": "x",
97
+ "56": "y",
98
+ "57": "z",
99
+ "58": "\u0410",
100
+ "59": "\u0411",
101
+ "60": "\u0412",
102
+ "61": "\u0413",
103
+ "62": "\u0414",
104
+ "63": "\u0415",
105
+ "64": "\u0416",
106
+ "65": "\u0417",
107
+ "66": "\u0418",
108
+ "67": "\u0419",
109
+ "68": "\u041a",
110
+ "69": "\u041b",
111
+ "70": "\u041c",
112
+ "71": "\u041d",
113
+ "72": "\u041e",
114
+ "73": "\u041f",
115
+ "74": "\u0420",
116
+ "75": "\u0421",
117
+ "76": "\u0422",
118
+ "77": "\u0423",
119
+ "78": "\u0424",
120
+ "79": "\u0425",
121
+ "80": "\u0426",
122
+ "81": "\u0427",
123
+ "82": "\u0428",
124
+ "83": "\u0429",
125
+ "84": "\u042b",
126
+ "85": "\u042c",
127
+ "86": "\u042d",
128
+ "87": "\u042e",
129
+ "88": "\u042f",
130
+ "89": "\u0430",
131
+ "90": "\u0431",
132
+ "91": "\u0432",
133
+ "92": "\u0433",
134
+ "93": "\u0434",
135
+ "94": "\u0435",
136
+ "95": "\u0436",
137
+ "96": "\u0437",
138
+ "97": "\u0438",
139
+ "98": "\u0439",
140
+ "99": "\u043a",
141
+ "100": "\u043b",
142
+ "101": "\u043c",
143
+ "102": "\u043d",
144
+ "103": "\u043e",
145
+ "104": "\u043f",
146
+ "105": "\u0440",
147
+ "106": "\u0441",
148
+ "107": "\u0442",
149
+ "108": "\u0443",
150
+ "109": "\u0444",
151
+ "110": "\u0445",
152
+ "111": "\u0446",
153
+ "112": "\u0447",
154
+ "113": "\u0448",
155
+ "114": "\u0449",
156
+ "115": "\u044a",
157
+ "116": "\u044b",
158
+ "117": "\u044c",
159
+ "118": "\u044d",
160
+ "119": "\u044e",
161
+ "120": "\u044f",
162
+ "121": "\u0451"
163
+ },
164
+ "init_std": 0.02,
165
+ "init_xavier_std": 1.0,
166
+ "is_encoder_decoder": true,
167
+ "label2id": {
168
+ " ": 5,
169
+ "A": 6,
170
+ "B": 7,
171
+ "C": 8,
172
+ "D": 9,
173
+ "E": 10,
174
+ "F": 11,
175
+ "G": 12,
176
+ "H": 13,
177
+ "I": 14,
178
+ "J": 15,
179
+ "K": 16,
180
+ "L": 17,
181
+ "M": 18,
182
+ "N": 19,
183
+ "O": 20,
184
+ "P": 21,
185
+ "Q": 22,
186
+ "R": 23,
187
+ "S": 24,
188
+ "T": 25,
189
+ "U": 26,
190
+ "V": 27,
191
+ "W": 28,
192
+ "X": 29,
193
+ "Y": 30,
194
+ "Z": 31,
195
+ "[CLS]": 2,
196
+ "[MASK]": 4,
197
+ "[PAD]": 0,
198
+ "[SEP]": 3,
199
+ "[UNK]": 1,
200
+ "a": 32,
201
+ "b": 33,
202
+ "c": 34,
203
+ "d": 35,
204
+ "e": 36,
205
+ "f": 37,
206
+ "g": 38,
207
+ "h": 39,
208
+ "i": 40,
209
+ "j": 41,
210
+ "k": 42,
211
+ "l": 43,
212
+ "m": 44,
213
+ "n": 45,
214
+ "o": 46,
215
+ "p": 47,
216
+ "q": 48,
217
+ "r": 49,
218
+ "s": 50,
219
+ "t": 51,
220
+ "u": 52,
221
+ "v": 53,
222
+ "w": 54,
223
+ "x": 55,
224
+ "y": 56,
225
+ "z": 57,
226
+ "\u0410": 58,
227
+ "\u0411": 59,
228
+ "\u0412": 60,
229
+ "\u0413": 61,
230
+ "\u0414": 62,
231
+ "\u0415": 63,
232
+ "\u0416": 64,
233
+ "\u0417": 65,
234
+ "\u0418": 66,
235
+ "\u0419": 67,
236
+ "\u041a": 68,
237
+ "\u041b": 69,
238
+ "\u041c": 70,
239
+ "\u041d": 71,
240
+ "\u041e": 72,
241
+ "\u041f": 73,
242
+ "\u0420": 74,
243
+ "\u0421": 75,
244
+ "\u0422": 76,
245
+ "\u0423": 77,
246
+ "\u0424": 78,
247
+ "\u0425": 79,
248
+ "\u0426": 80,
249
+ "\u0427": 81,
250
+ "\u0428": 82,
251
+ "\u0429": 83,
252
+ "\u042b": 84,
253
+ "\u042c": 85,
254
+ "\u042d": 86,
255
+ "\u042e": 87,
256
+ "\u042f": 88,
257
+ "\u0430": 89,
258
+ "\u0431": 90,
259
+ "\u0432": 91,
260
+ "\u0433": 92,
261
+ "\u0434": 93,
262
+ "\u0435": 94,
263
+ "\u0436": 95,
264
+ "\u0437": 96,
265
+ "\u0438": 97,
266
+ "\u0439": 98,
267
+ "\u043a": 99,
268
+ "\u043b": 100,
269
+ "\u043c": 101,
270
+ "\u043d": 102,
271
+ "\u043e": 103,
272
+ "\u043f": 104,
273
+ "\u0440": 105,
274
+ "\u0441": 106,
275
+ "\u0442": 107,
276
+ "\u0443": 108,
277
+ "\u0444": 109,
278
+ "\u0445": 110,
279
+ "\u0446": 111,
280
+ "\u0447": 112,
281
+ "\u0448": 113,
282
+ "\u0449": 114,
283
+ "\u044a": 115,
284
+ "\u044b": 116,
285
+ "\u044c": 117,
286
+ "\u044d": 118,
287
+ "\u044e": 119,
288
+ "\u044f": 120,
289
+ "\u0451": 121
290
+ },
291
+ "mask_loss_coefficient": 1,
292
+ "max_position_embeddings": 1024,
293
+ "model_type": "detr",
294
+ "num_channels": 3,
295
+ "num_hidden_layers": 6,
296
+ "num_queries": 100,
297
+ "position_embedding_type": "sine",
298
+ "scale_embedding": false,
299
+ "torch_dtype": "float32",
300
+ "transformers_version": "4.47.0",
301
+ "use_pretrained_backbone": true,
302
+ "use_timm_backbone": true
303
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:47d433c3abbe3e1e3320bba089f052e2169e7892a4dc3fbddfb9acc57a5cc28b
3
+ size 166619244