{ "sae_lens_version": "4.0.0", "tokenizer_name": "NeelNanda/gpt-neox-tokenizer-digits", "original_dataset": "NeelNanda/c4-10k", "original_split": "train[:100]", "original_data_files": null, "context_size": 16, "shuffled": true, "seed": null, "begin_batch_token": "bos", "begin_sequence_token": null, "sequence_separator_token": "eos" }