jerryzh168 commited on
Commit
ffa4b82
·
verified ·
1 Parent(s): 7559662

Upload Phi3ForCausalLM

Browse files
Files changed (3) hide show
  1. config.json +15 -9
  2. generation_config.json +1 -1
  3. pytorch_model.bin +2 -2
config.json CHANGED
@@ -5,9 +5,9 @@
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "auto_map": {
8
- "AutoConfig": "microsoft/Phi-4-mini-instruct--configuration_phi3.Phi3Config",
9
- "AutoModelForCausalLM": "microsoft/Phi-4-mini-instruct--modeling_phi3.Phi3ForCausalLM",
10
- "AutoTokenizer": "microsoft/Phi-4-mini-instruct--Xenova/gpt-4o"
11
  },
12
  "bos_token_id": 199999,
13
  "embd_pdrop": 0.0,
@@ -29,6 +29,7 @@
29
  "pad_token_id": 199999,
30
  "partial_rotary_factor": 0.75,
31
  "quantization_config": {
 
32
  "modules_to_not_convert": null,
33
  "quant_method": "torchao",
34
  "quant_type": {
@@ -37,6 +38,14 @@
37
  "base_config": {
38
  "_data": {
39
  "group_size": 128,
 
 
 
 
 
 
 
 
40
  "layout": {
41
  "_data": {
42
  "inner_k_tiles": 8
@@ -44,10 +53,6 @@
44
  "_type": "TensorCoreTiledLayout",
45
  "_version": 1
46
  },
47
- "packing_format": {
48
- "_data": "PLAIN",
49
- "_type": "PackingFormat"
50
- },
51
  "preserve_zero": null,
52
  "set_inductor_config": true,
53
  "use_hqq": false,
@@ -66,7 +71,8 @@
66
  "_version": 1
67
  }
68
  },
69
- "quant_type_kwargs": {}
 
70
  },
71
  "resid_pdrop": 0.0,
72
  "rms_norm_eps": 1e-05,
@@ -177,7 +183,7 @@
177
  "sliding_window": 262144,
178
  "tie_word_embeddings": true,
179
  "torch_dtype": "bfloat16",
180
- "transformers_version": "4.52.0.dev0",
181
  "use_cache": true,
182
  "vocab_size": 200064
183
  }
 
5
  "attention_bias": false,
6
  "attention_dropout": 0.0,
7
  "auto_map": {
8
+ "AutoConfig": "configuration_phi3.Phi3Config",
9
+ "AutoModelForCausalLM": "modeling_phi3.Phi3ForCausalLM",
10
+ "AutoTokenizer": "Xenova/gpt-4o"
11
  },
12
  "bos_token_id": 199999,
13
  "embd_pdrop": 0.0,
 
29
  "pad_token_id": 199999,
30
  "partial_rotary_factor": 0.75,
31
  "quantization_config": {
32
+ "include_input_output_embeddings": false,
33
  "modules_to_not_convert": null,
34
  "quant_method": "torchao",
35
  "quant_type": {
 
38
  "base_config": {
39
  "_data": {
40
  "group_size": 128,
41
+ "int4_choose_qparams_algorithm": {
42
+ "_data": "TINYGEMM",
43
+ "_type": "Int4ChooseQParamsAlgorithm"
44
+ },
45
+ "int4_packing_format": {
46
+ "_data": "PLAIN",
47
+ "_type": "Int4PackingFormat"
48
+ },
49
  "layout": {
50
  "_data": {
51
  "inner_k_tiles": 8
 
53
  "_type": "TensorCoreTiledLayout",
54
  "_version": 1
55
  },
 
 
 
 
56
  "preserve_zero": null,
57
  "set_inductor_config": true,
58
  "use_hqq": false,
 
71
  "_version": 1
72
  }
73
  },
74
+ "quant_type_kwargs": {},
75
+ "untie_embedding_weights": false
76
  },
77
  "resid_pdrop": 0.0,
78
  "rms_norm_eps": 1e-05,
 
183
  "sliding_window": 262144,
184
  "tie_word_embeddings": true,
185
  "torch_dtype": "bfloat16",
186
+ "transformers_version": "4.55.4",
187
  "use_cache": true,
188
  "vocab_size": 200064
189
  }
generation_config.json CHANGED
@@ -6,5 +6,5 @@
6
  199999
7
  ],
8
  "pad_token_id": 199999,
9
- "transformers_version": "4.52.0.dev0"
10
  }
 
6
  199999
7
  ],
8
  "pad_token_id": 199999,
9
+ "transformers_version": "4.55.4"
10
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e2a528453e5a1be0710b2a8b22331566d3e4b9361a04f60dd4be8570eaf04d08
3
- size 3268699771
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:007dccc15cb8fe1198f6155f338598b84fa4971e4562f31bfb82632a0fc692d6
3
+ size 3268687035