Xenova HF Staff commited on
Commit
8da63a2
·
verified ·
1 Parent(s): 1bd2d2a

Update config.json

Browse files
Files changed (1) hide show
  1. config.json +1 -27
config.json CHANGED
@@ -12,37 +12,11 @@
12
  "initializer_range": 0.02,
13
  "intermediate_size": 32,
14
  "layer_types": [
15
- "full_attention",
16
- "full_attention",
17
- "full_attention",
18
- "full_attention",
19
- "full_attention",
20
- "full_attention",
21
- "full_attention",
22
- "full_attention",
23
- "full_attention",
24
- "full_attention",
25
- "full_attention",
26
- "full_attention",
27
- "full_attention",
28
- "full_attention",
29
- "full_attention",
30
- "full_attention",
31
- "full_attention",
32
- "full_attention",
33
- "full_attention",
34
- "full_attention",
35
- "full_attention",
36
- "full_attention",
37
- "full_attention",
38
- "full_attention",
39
- "full_attention",
40
- "full_attention",
41
  "full_attention",
42
  "full_attention"
43
  ],
44
  "max_position_embeddings": 40960,
45
- "max_window_layers": 28,
46
  "model_type": "qwen3",
47
  "num_attention_heads": 4,
48
  "num_hidden_layers": 2,
 
12
  "initializer_range": 0.02,
13
  "intermediate_size": 32,
14
  "layer_types": [
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  "full_attention",
16
  "full_attention"
17
  ],
18
  "max_position_embeddings": 40960,
19
+ "max_window_layers": 2,
20
  "model_type": "qwen3",
21
  "num_attention_heads": 4,
22
  "num_hidden_layers": 2,