{ | |
"architectures": ["SPTForCausalLM"], | |
"model_type": "spt", | |
"n_layers": 12, | |
"vocab_size": 97, | |
"hidden_size": 512, | |
"n_attn_heads": 16, | |
"n_kv_heads": 16, | |
"intermediate_size": 2048, | |
"max_len": 2048, | |
"residual": true, | |
"normalise": true, | |
"bos_token_id": 95, | |
"eos_token_id": 95, | |
"pad_token_id": 95, | |
"unk_token_id": 96 | |
} |