--- configs: - config_name: default data_files: - split: train path: data/train-* - split: validation path: data/validation-* - split: test path: data/test-* dataset_info: features: - name: text dtype: string splits: - name: train num_bytes: 1136718026.846949 num_examples: 452484 - name: validation num_bytes: 30473651.26394911 num_examples: 11908 - name: test num_bytes: 30471237.904544305 num_examples: 11908 download_size: 538623929 dataset_size: 1197662916.0154426 source_datasets: - pszemraj/simple_wikipedia_LM - pszemraj/refinedweb-3m-deduped-split --- # Dataset Card for "simpleRW-lite" interleaved simple wikipedia LM + refinedweb-3m ```python DatasetDict({ train: Dataset({ features: ['text'], num_rows: 452484 }) validation: Dataset({ features: ['text'], num_rows: 11908 }) test: Dataset({ features: ['text'], num_rows: 11908 }) }) ``` train ``` Descriptive Stats Using Pandas: count 452484.000000 mean 430.923633 std 1391.959655 min 0.000000 25% 83.000000 50% 175.000000 75% 432.000000 max 135922.000000 dtype: float64 ``` [More Information needed](https://github.com/huggingface/datasets/blob/main/CONTRIBUTING.md#how-to-contribute-to-the-dataset-cards)