{ "architectures": [ "ByteGPTForCausalLM" ], "auto_map": { "AutoConfig": "configuration_bytegpt.ByteGPTConfig", "AutoModelForCausalLM": "modeling_bytegpt.ByteGPTForCausalLM" }, "block_size": 1024, "dropout": 0.1, "model_type": "ijk_byte_gpt", "n_embd": 768, "n_head": 12, "n_layer": 12, "torch_dtype": "float32", "transformers_version": "4.48.2", "use_flash_attention": false, "vocab_size": 256 }