lysandre HF Staff commited on
Commit
03d0524
·
1 Parent(s): 5ed2bc1

Upload dataset_infos.json with huggingface_hub

Browse files
Files changed (1) hide show
  1. dataset_infos.json +28 -28
dataset_infos.json CHANGED
@@ -14,12 +14,6 @@
14
  }
15
  },
16
  "splits": {
17
- "gradio": {
18
- "name": "gradio",
19
- "num_bytes": 20526,
20
- "num_examples": 933,
21
- "dataset_name": null
22
- },
23
  "diffusers": {
24
  "name": "diffusers",
25
  "num_bytes": 6534,
@@ -32,56 +26,62 @@
32
  "num_examples": 415,
33
  "dataset_name": null
34
  },
35
- "transformers": {
36
- "name": "transformers",
37
- "num_bytes": 21208,
38
- "num_examples": 964,
39
- "dataset_name": null
40
- },
41
- "accelerate": {
42
- "name": "accelerate",
43
- "num_bytes": 17160,
44
- "num_examples": 780,
45
- "dataset_name": null
46
- },
47
  "huggingface_hub": {
48
  "name": "huggingface_hub",
49
  "num_bytes": 18040,
50
  "num_examples": 820,
51
  "dataset_name": null
52
  },
53
- "datasets": {
54
- "name": "datasets",
55
- "num_bytes": 17160,
56
- "num_examples": 780,
57
- "dataset_name": null
58
- },
59
  "optimum": {
60
  "name": "optimum",
61
  "num_bytes": 12144,
62
  "num_examples": 552,
63
  "dataset_name": null
64
  },
 
 
 
 
 
 
65
  "tokenizers": {
66
  "name": "tokenizers",
67
  "num_bytes": 20526,
68
  "num_examples": 933,
69
  "dataset_name": null
70
  },
 
 
 
 
 
 
71
  "peft": {
72
  "name": "peft",
73
  "num_bytes": 1386,
74
  "num_examples": 63,
75
  "dataset_name": null
76
  },
77
- "pytorch_image_models": {
78
- "name": "pytorch_image_models",
 
 
 
 
 
 
79
  "num_bytes": 20526,
80
  "num_examples": 933,
81
  "dataset_name": null
 
 
 
 
 
 
82
  }
83
  },
84
- "download_size": 96245,
85
  "dataset_size": 164340,
86
- "size_in_bytes": 260585
87
  }}
 
14
  }
15
  },
16
  "splits": {
 
 
 
 
 
 
17
  "diffusers": {
18
  "name": "diffusers",
19
  "num_bytes": 6534,
 
26
  "num_examples": 415,
27
  "dataset_name": null
28
  },
 
 
 
 
 
 
 
 
 
 
 
 
29
  "huggingface_hub": {
30
  "name": "huggingface_hub",
31
  "num_bytes": 18040,
32
  "num_examples": 820,
33
  "dataset_name": null
34
  },
 
 
 
 
 
 
35
  "optimum": {
36
  "name": "optimum",
37
  "num_bytes": 12144,
38
  "num_examples": 552,
39
  "dataset_name": null
40
  },
41
+ "pytorch_image_models": {
42
+ "name": "pytorch_image_models",
43
+ "num_bytes": 20526,
44
+ "num_examples": 933,
45
+ "dataset_name": null
46
+ },
47
  "tokenizers": {
48
  "name": "tokenizers",
49
  "num_bytes": 20526,
50
  "num_examples": 933,
51
  "dataset_name": null
52
  },
53
+ "accelerate": {
54
+ "name": "accelerate",
55
+ "num_bytes": 17160,
56
+ "num_examples": 780,
57
+ "dataset_name": null
58
+ },
59
  "peft": {
60
  "name": "peft",
61
  "num_bytes": 1386,
62
  "num_examples": 63,
63
  "dataset_name": null
64
  },
65
+ "datasets": {
66
+ "name": "datasets",
67
+ "num_bytes": 17160,
68
+ "num_examples": 780,
69
+ "dataset_name": null
70
+ },
71
+ "gradio": {
72
+ "name": "gradio",
73
  "num_bytes": 20526,
74
  "num_examples": 933,
75
  "dataset_name": null
76
+ },
77
+ "transformers": {
78
+ "name": "transformers",
79
+ "num_bytes": 21208,
80
+ "num_examples": 964,
81
+ "dataset_name": null
82
  }
83
  },
84
+ "download_size": 0,
85
  "dataset_size": 164340,
86
+ "size_in_bytes": 164340
87
  }}