Datasets:

Formats:
parquet
Libraries:
Datasets
Dask
License:
jwkirchenbauer commited on
Commit
88f9c11
·
verified ·
1 Parent(s): 9eb8d14

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +18 -1
README.md CHANGED
@@ -5,4 +5,21 @@ configs:
5
  - split: train
6
  path: "*.parquet"
7
  ---
8
- Gemstones Training Dataset - Worker sharded version
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  - split: train
6
  path: "*.parquet"
7
  ---
8
+ Gemstones Training Dataset - Worker sharded version
9
+
10
+ **Disclaimer:** this is an approximation of the dataset used to train the Gemstones model suite.
11
+ Due to the randomized and sharded nature of the distributed training code, the only way to perfectly
12
+ reproduce the training bactches across the gpus is/was the run the training code.
13
+ This is the result of an attempt to simulate the way in which the training code loaded the data and
14
+ stream it out to a portable file format for use in downstream analyses of the model suite.
15
+
16
+ # Sharding format: worker parallel
17
+
18
+ This version of th
19
+
20
+
21
+ # Loading
22
+
23
+ This data should be loadable using `load_dataset` in the standard manner to auto-download the data.
24
+ Alternately, the dataset can be cloned using git to materialize the files locally, and then loaded
25
+ using the default `parquet` builder as described here: https://huggingface.co/datasets/tomg-group-umd/gemstones_data_order_parallel