omeryagmurlu commited on
Commit
b64c1c0
·
verified ·
1 Parent(s): b6cdb49

Add files using upload-large-folder tool

Browse files
.gitattributes CHANGED
@@ -57,3 +57,19 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
57
  # Video files - compressed
58
  *.mp4 filter=lfs diff=lfs merge=lfs -text
59
  *.webm filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  # Video files - compressed
58
  *.mp4 filter=lfs diff=lfs merge=lfs -text
59
  *.webm filter=lfs diff=lfs merge=lfs -text
60
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00010-of-00016 filter=lfs diff=lfs merge=lfs -text
61
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00014-of-00016 filter=lfs diff=lfs merge=lfs -text
62
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00002-of-00016 filter=lfs diff=lfs merge=lfs -text
63
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00013-of-00016 filter=lfs diff=lfs merge=lfs -text
64
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00005-of-00016 filter=lfs diff=lfs merge=lfs -text
65
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00000-of-00016 filter=lfs diff=lfs merge=lfs -text
66
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00012-of-00016 filter=lfs diff=lfs merge=lfs -text
67
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00006-of-00016 filter=lfs diff=lfs merge=lfs -text
68
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00004-of-00016 filter=lfs diff=lfs merge=lfs -text
69
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00008-of-00016 filter=lfs diff=lfs merge=lfs -text
70
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00001-of-00016 filter=lfs diff=lfs merge=lfs -text
71
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00009-of-00016 filter=lfs diff=lfs merge=lfs -text
72
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00015-of-00016 filter=lfs diff=lfs merge=lfs -text
73
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00011-of-00016 filter=lfs diff=lfs merge=lfs -text
74
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00003-of-00016 filter=lfs diff=lfs merge=lfs -text
75
+ 1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00007-of-00016 filter=lfs diff=lfs merge=lfs -text
1.0.0/README.MD ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ Dataset with joint state play data for a toy kitchen.
2
+ The dataset contains 20 different tasks with two static cameras.
3
+
4
+ @inproceedings{
5
+ reuss2024multimodal,
6
+ title={Multimodal Diffusion Transformer: Learning Versatile Behavior from Multimodal Goals},
7
+ author={Moritz Reuss and {\"O}mer Erdin{\c{c}} Ya{\u{g}}murlu and Fabian Wenzel and Rudolf Lioutikov},
8
+ booktitle={Robotics: Science and Systems},
9
+ year={2024}
10
+ }
1.0.0/dataset_info.json ADDED
@@ -0,0 +1,36 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "citation": "// TODO(example_dataset): BibTeX citation",
3
+ "description": "TODO(example_dataset): Markdown description of your dataset.\nDescription is **formatted** as markdown.\n\nIt should also contain any processing which has been applied (if any),\n(e.g. corrupted example skipped, images cropped,...):",
4
+ "fileFormat": "tfrecord",
5
+ "moduleName": "kit_irl_real_kitchen_lang.kit_irl_real_kitchen_lang",
6
+ "name": "kit_irl_real_kitchen_lang",
7
+ "releaseNotes": {
8
+ "1.0.0": "Initial release."
9
+ },
10
+ "splits": [
11
+ {
12
+ "filepathTemplate": "{DATASET}-{SPLIT}.{FILEFORMAT}-{SHARD_X_OF_Y}",
13
+ "name": "train",
14
+ "numBytes": "1555997050",
15
+ "shardLengths": [
16
+ "26",
17
+ "26",
18
+ "26",
19
+ "26",
20
+ "26",
21
+ "26",
22
+ "26",
23
+ "26",
24
+ "27",
25
+ "26",
26
+ "26",
27
+ "26",
28
+ "26",
29
+ "26",
30
+ "26",
31
+ "26"
32
+ ]
33
+ }
34
+ ],
35
+ "version": "1.0.0"
36
+ }
1.0.0/dataset_statistics_cfcf3bdcc3f95b1d539a24749ade2a7e36ff0ba1e66407e4610cdaa129e4c10f.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"action": {"mean": [0.2703530788421631, 0.1182452067732811, 0.35498133301734924, -1.5729808807373047, -0.5235288739204407, 1.9427248239517212, -1.5018268823623657, 0.5981078743934631], "std": [0.12345685809850693, 0.3061469495296478, 0.27789995074272156, 0.4563933312892914, 0.5890739560127258, 0.30609846115112305, 0.44644784927368164, 0.4899592697620392], "max": [0.6685911417007446, 1.0151987075805664, 1.3512533903121948, -0.3098829984664917, 0.15463855862617493, 3.230781078338623, 0.4125751256942749, 1.0], "min": [-0.4728304147720337, -0.9769639372825623, -0.3555123209953308, -2.877141237258911, -2.5706896781921387, 0.6824776530265808, -2.8376104831695557, 0.0], "p99": [0.5422249287366867, 0.8028479218482971, 1.0687870979309082, -0.9487988352775574, 0.06975962594151497, 2.624782383441925, -0.49846622347831726, 1.0], "p01": [-0.08110086619853973, -0.6993335634469986, -0.17626342177391052, -2.7020722031593323, -2.432827055454254, 1.232361376285553, -2.728285014629364, 0.0]}, "num_transitions": 25276, "num_trajectories": 417, "proprio": {"mean": [0.2709193229675293, 0.12117421627044678, 0.3550775647163391, -1.576100468635559, -0.5228900909423828, 1.9408499002456665, -1.5027081966400146, 0.5981078743934631], "std": [0.12378045171499252, 0.30774423480033875, 0.27788597345352173, 0.456500381231308, 0.5882568955421448, 0.304630309343338, 0.4435570538043976, 0.4899592697620392], "max": [0.6807584166526794, 1.02162504196167, 1.3513909578323364, -0.3086922764778137, 0.14923511445522308, 3.2253408432006836, 0.4084588587284088, 1.0], "min": [-0.472444623708725, -0.9860996603965759, -0.3634328246116638, -2.8732833862304688, -2.5652005672454834, 0.6849313974380493, -2.787592887878418, 0.0], "p99": [0.5407610535621643, 0.807393342256546, 1.0714323222637177, -0.9510097801685333, 0.0711765754967928, 2.6280637383461, -0.5043639838695526, 1.0], "p01": [-0.08156387694180012, -0.7015249878168106, -0.17795844748616219, -2.704028069972992, -2.4300113916397095, 1.2408689856529236, -2.723939895629883, 0.0]}}
1.0.0/features.json ADDED
@@ -0,0 +1,291 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict",
3
+ "featuresDict": {
4
+ "features": {
5
+ "steps": {
6
+ "pythonClassName": "tensorflow_datasets.core.features.dataset_feature.Dataset",
7
+ "sequence": {
8
+ "feature": {
9
+ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict",
10
+ "featuresDict": {
11
+ "features": {
12
+ "action_abs": {
13
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
14
+ "tensor": {
15
+ "shape": {
16
+ "dimensions": [
17
+ "7"
18
+ ]
19
+ },
20
+ "dtype": "float64",
21
+ "encoding": "none"
22
+ },
23
+ "description": "Absolute robot action, consists of [3x delta_end_effector_pos, 3x delta_end_effector_ori (euler: roll, pitch, yaw), 1x des_gripper_width]."
24
+ },
25
+ "action_joint_state": {
26
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
27
+ "tensor": {
28
+ "shape": {
29
+ "dimensions": [
30
+ "7"
31
+ ]
32
+ },
33
+ "dtype": "float64",
34
+ "encoding": "none"
35
+ },
36
+ "description": "Robot action in joint space, consists of [7x joint states]"
37
+ },
38
+ "is_terminal": {
39
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
40
+ "tensor": {
41
+ "shape": {},
42
+ "dtype": "bool",
43
+ "encoding": "none"
44
+ },
45
+ "description": "True on last step of the episode if it is a terminal step, True for demos."
46
+ },
47
+ "is_last": {
48
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
49
+ "tensor": {
50
+ "shape": {},
51
+ "dtype": "bool",
52
+ "encoding": "none"
53
+ },
54
+ "description": "True on last step of the episode."
55
+ },
56
+ "language_instruction_3": {
57
+ "pythonClassName": "tensorflow_datasets.core.features.text_feature.Text",
58
+ "text": {},
59
+ "description": "Language Instruction."
60
+ },
61
+ "observation": {
62
+ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict",
63
+ "featuresDict": {
64
+ "features": {
65
+ "end_effector_ori": {
66
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
67
+ "tensor": {
68
+ "shape": {
69
+ "dimensions": [
70
+ "3"
71
+ ]
72
+ },
73
+ "dtype": "float64",
74
+ "encoding": "none"
75
+ },
76
+ "description": "Current End Effector orientation in Cartesian space as Euler (xyz)"
77
+ },
78
+ "image_top": {
79
+ "pythonClassName": "tensorflow_datasets.core.features.image_feature.Image",
80
+ "image": {
81
+ "shape": {
82
+ "dimensions": [
83
+ "250",
84
+ "250",
85
+ "3"
86
+ ]
87
+ },
88
+ "dtype": "uint8",
89
+ "encodingFormat": "jpeg"
90
+ },
91
+ "description": "Main camera RGB observation."
92
+ },
93
+ "end_effector_ori_quat": {
94
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
95
+ "tensor": {
96
+ "shape": {
97
+ "dimensions": [
98
+ "4"
99
+ ]
100
+ },
101
+ "dtype": "float64",
102
+ "encoding": "none"
103
+ },
104
+ "description": "Current End Effector orientation in Cartesian space as Quaternion"
105
+ },
106
+ "end_effector_pos": {
107
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
108
+ "tensor": {
109
+ "shape": {
110
+ "dimensions": [
111
+ "3"
112
+ ]
113
+ },
114
+ "dtype": "float64",
115
+ "encoding": "none"
116
+ },
117
+ "description": "Current End Effector position in Cartesian space"
118
+ },
119
+ "joint_state": {
120
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
121
+ "tensor": {
122
+ "shape": {
123
+ "dimensions": [
124
+ "7"
125
+ ]
126
+ },
127
+ "dtype": "float64",
128
+ "encoding": "none"
129
+ },
130
+ "description": "Robot joint state. Consists of [7x joint states]"
131
+ },
132
+ "image_side": {
133
+ "pythonClassName": "tensorflow_datasets.core.features.image_feature.Image",
134
+ "image": {
135
+ "shape": {
136
+ "dimensions": [
137
+ "250",
138
+ "250",
139
+ "3"
140
+ ]
141
+ },
142
+ "dtype": "uint8",
143
+ "encodingFormat": "jpeg"
144
+ },
145
+ "description": "Wrist camera RGB observation."
146
+ },
147
+ "joint_state_velocity": {
148
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
149
+ "tensor": {
150
+ "shape": {
151
+ "dimensions": [
152
+ "7"
153
+ ]
154
+ },
155
+ "dtype": "float64",
156
+ "encoding": "none"
157
+ },
158
+ "description": "Robot joint velocities. Consists of [7x joint velocities]"
159
+ }
160
+ }
161
+ }
162
+ },
163
+ "discount": {
164
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
165
+ "tensor": {
166
+ "shape": {},
167
+ "dtype": "float64",
168
+ "encoding": "none"
169
+ },
170
+ "description": "Discount if provided, default to 1."
171
+ },
172
+ "action_joint_vel": {
173
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
174
+ "tensor": {
175
+ "shape": {
176
+ "dimensions": [
177
+ "7"
178
+ ]
179
+ },
180
+ "dtype": "float64",
181
+ "encoding": "none"
182
+ },
183
+ "description": "Robot action in joint space, consists of [7x joint velocities]"
184
+ },
185
+ "reward": {
186
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
187
+ "tensor": {
188
+ "shape": {},
189
+ "dtype": "float64",
190
+ "encoding": "none"
191
+ },
192
+ "description": "Reward if provided, 1 on final step for demos."
193
+ },
194
+ "language_embedding": {
195
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
196
+ "tensor": {
197
+ "shape": {
198
+ "dimensions": [
199
+ "3",
200
+ "512"
201
+ ]
202
+ },
203
+ "dtype": "float32",
204
+ "encoding": "none"
205
+ },
206
+ "description": "Kona language embedding. See https://tfhub.dev/google/universal-sentence-encoder-large/5"
207
+ },
208
+ "language_instruction_2": {
209
+ "pythonClassName": "tensorflow_datasets.core.features.text_feature.Text",
210
+ "text": {},
211
+ "description": "Language Instruction."
212
+ },
213
+ "language_instruction": {
214
+ "pythonClassName": "tensorflow_datasets.core.features.text_feature.Text",
215
+ "text": {},
216
+ "description": "Language Instruction."
217
+ },
218
+ "is_first": {
219
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
220
+ "tensor": {
221
+ "shape": {},
222
+ "dtype": "bool",
223
+ "encoding": "none"
224
+ },
225
+ "description": "True on first step of the episode."
226
+ },
227
+ "action_gripper_width": {
228
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
229
+ "tensor": {
230
+ "shape": {},
231
+ "dtype": "float64",
232
+ "encoding": "none"
233
+ },
234
+ "description": "Desired gripper width, consists of [1x gripper width] in range [0, 1]"
235
+ },
236
+ "delta_des_joint_state": {
237
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
238
+ "tensor": {
239
+ "shape": {
240
+ "dimensions": [
241
+ "7"
242
+ ]
243
+ },
244
+ "dtype": "float64",
245
+ "encoding": "none"
246
+ },
247
+ "description": "Delta robot action in joint space, consists of [7x joint states]"
248
+ },
249
+ "action": {
250
+ "pythonClassName": "tensorflow_datasets.core.features.tensor_feature.Tensor",
251
+ "tensor": {
252
+ "shape": {
253
+ "dimensions": [
254
+ "7"
255
+ ]
256
+ },
257
+ "dtype": "float64",
258
+ "encoding": "none"
259
+ },
260
+ "description": "Delta robot action, consists of [3x delta_end_effector_pos, 3x delta_end_effector_ori (euler: roll, pitch, yaw), 1x des_gripper_width]."
261
+ }
262
+ }
263
+ }
264
+ },
265
+ "length": "-1"
266
+ }
267
+ },
268
+ "episode_metadata": {
269
+ "pythonClassName": "tensorflow_datasets.core.features.features_dict.FeaturesDict",
270
+ "featuresDict": {
271
+ "features": {
272
+ "traj_length": {
273
+ "pythonClassName": "tensorflow_datasets.core.features.scalar.Scalar",
274
+ "tensor": {
275
+ "shape": {},
276
+ "dtype": "float64",
277
+ "encoding": "none"
278
+ },
279
+ "description": "Number of samples in trajectorie"
280
+ },
281
+ "file_path": {
282
+ "pythonClassName": "tensorflow_datasets.core.features.text_feature.Text",
283
+ "text": {},
284
+ "description": "Path to the original data file."
285
+ }
286
+ }
287
+ }
288
+ }
289
+ }
290
+ }
291
+ }
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00000-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d9f5281500daa8b76b8efc41659461fc32284a1b5746bcc2affbeea4f12b7fe
3
+ size 95775764
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00001-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7be02cd7a37f89c67f94d6f7a0ca1c6aaf00e21c8795ecf225be80b0c2718469
3
+ size 96207287
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00002-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0598ab269b0e5a0d8a64d6adf8f28fc16d6eb20b42b28d0864c67523cf9c799c
3
+ size 100410465
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00003-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6247c9ca76de972c17e46f95a458b05208fc35991c61cdf3cbbb2d3cffc2d1bc
3
+ size 99494307
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00004-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ed950899e488efbf37a4a9a63bd52bfe4bec5733fad5ced51654f0de09cfc06
3
+ size 93715351
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00005-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac489217058196ee8fd79d45700a5ffb3b407db4ec93e8a7cf6a4f8309d5c417
3
+ size 92129037
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00006-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b8100538cd096578510fcbe4ae83e7ab2f484d17806d4aa918974dcba2066c64
3
+ size 93580398
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00007-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09fbdb20a39b8256884f047967be259649df7b5618aadcd8b3cd9b4468635068
3
+ size 99562556
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00008-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1a24291b3ae77e71045d8f6375aa313ddb71d472ce965ad327d9ee8008314d46
3
+ size 100747564
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00009-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba651073435f3b55dbd82adeb8cb71ffccac3fbe8ccf1be489570a60868ee2c6
3
+ size 102437793
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00010-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:207d33d23926261ef73d0e3676c63ba18572f2e87ad2ea63275732d07bec505d
3
+ size 93943701
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00011-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c8384a3fda3553ba5fb05b12cc3966d4d8d158363257bcb97086500d1c49004
3
+ size 97720723
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00012-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96892b7364aa71ab41b5fe0abe64adb8d22377db62a6bc704a30b5ff22c68ae5
3
+ size 89658168
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00013-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ad4aeadeee52a291a24f58fc2ff0ccdfda59458d0859438d54e55bd1f0aa6fd6
3
+ size 94680480
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00014-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2b23fc5ae1fa937602630bdc40600b54db0849fb2f1ef2501a654bdf0cc161df
3
+ size 94454238
1.0.0/kit_irl_real_kitchen_lang-train.tfrecord-00015-of-00016 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2a252beeeb5f1de2ac5227d6810137519419b73125cab34c54dd37fe5b3b30c
3
+ size 111485890