zhujiangang commited on
Commit
227f47c
·
verified ·
1 Parent(s): 019becf

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +10 -5
config.json CHANGED
@@ -21,13 +21,17 @@
21
  "num_experts": 64,
22
  "num_shared_experts": 1,
23
  "norm_topk_prob": true,
24
- "num_attention_heads": 42,
25
  "num_experts_per_tok": 4,
26
  "num_hidden_layers": 88,
27
- "num_key_value_heads": 6,
28
  "pretraining_tp": 1,
29
- "rms_norm_eps": 1e-06,
30
- "rope_scaling": null,
 
 
 
 
31
  "rope_theta": 600000,
32
  "tie_word_embeddings": false,
33
  "torch_dtype": "bfloat16",
@@ -40,5 +44,6 @@
40
  "embedding_dropout": 0.0,
41
  "norm_head": true,
42
  "norm_softmax": false,
43
- "output_dropout": 0.0
 
44
  }
 
21
  "num_experts": 64,
22
  "num_shared_experts": 1,
23
  "norm_topk_prob": true,
24
+ "num_attention_heads": 56,
25
  "num_experts_per_tok": 4,
26
  "num_hidden_layers": 88,
27
+ "num_key_value_heads": 8,
28
  "pretraining_tp": 1,
29
+ "rms_norm_eps": 1e-05,
30
+ "rope_scaling": {
31
+ "factor": 4.0,
32
+ "original_max_position_embeddings": 16384,
33
+ "type": "yarn"
34
+ },
35
  "rope_theta": 600000,
36
  "tie_word_embeddings": false,
37
  "torch_dtype": "bfloat16",
 
44
  "embedding_dropout": 0.0,
45
  "norm_head": true,
46
  "norm_softmax": false,
47
+ "output_dropout": 0.0,
48
+ "head_dim": 128
49
  }