File size: 418 Bytes
9aa8ed3 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 |
from dataclasses import dataclass
@dataclass
class LlamaConfig:
hidden_size: int = 576
num_attention_heads: int = 9
num_key_value_heads: int = 3
num_hidden_layers: int = 30
intermediate_size: int = 1536
hidden_act: str = "silu"
rms_norm_eps: float = 1e-5
vocab_size: int = 49152
max_position_embeddings: int = 8192
rope_theta: int = 100000
tie_word_embeddings: bool = False |