sharpenb commited on
Commit
c8ace31
·
verified ·
1 Parent(s): 7fe99be

Upload folder using huggingface_hub (#5)

Browse files

- 76a9f3229c0d1aa7d577eb3a45e8a2df9ad9fc14c060b645c91913c43be12490 (7854c0c16a77587f06d94514342041ed47a93c7c)
- f2accc381628497ef754aa39324aac57145c0574448f042e051887b91707563d (23988bca5ad1f87ff0e70e471afb3c9cdd23ebc5)

Files changed (3) hide show
  1. config.json +21 -1
  2. qmodel.pt +2 -2
  3. smash_config.json +1 -1
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "_attn_implementation_autoset": true,
3
- "_name_or_path": "HuggingFaceTB/SmolLM2-1.7B-Instruct",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
@@ -21,6 +21,26 @@
21
  "num_key_value_heads": 32,
22
  "pad_token_id": 2,
23
  "pretraining_tp": 1,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  "rms_norm_eps": 1e-05,
25
  "rope_scaling": null,
26
  "rope_theta": 130000,
 
1
  {
2
  "_attn_implementation_autoset": true,
3
+ "_name_or_path": "/tmp/models/tmpuszyy52l/tmpu5eb0hd8",
4
  "architectures": [
5
  "LlamaForCausalLM"
6
  ],
 
21
  "num_key_value_heads": 32,
22
  "pad_token_id": 2,
23
  "pretraining_tp": 1,
24
+ "quantization_config": {
25
+ "quant_config": {
26
+ "offload_meta": false,
27
+ "scale_quant_params": null,
28
+ "weight_quant_params": {
29
+ "axis": 1,
30
+ "channel_wise": true,
31
+ "group_size": 64,
32
+ "nbits": 8,
33
+ "optimize": true,
34
+ "round_zero": false,
35
+ "view_as_float": false
36
+ },
37
+ "zero_quant_params": null
38
+ },
39
+ "quant_method": "hqq",
40
+ "skip_modules": [
41
+ "lm_head"
42
+ ]
43
+ },
44
  "rms_norm_eps": 1e-05,
45
  "rope_scaling": null,
46
  "rope_theta": 130000,
qmodel.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:81cd9322efd0bf8fd087d644232db97b9a282b56ee59f97c7ddbd81951cee622
3
- size 1912983472
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d54f27b8eec5c7e905958321aee5898dae3a5223ee7775ca2fba42ffd20cd1a0
3
+ size 2215166128
smash_config.json CHANGED
@@ -11,7 +11,7 @@
11
  "quant_hqq_weight_bits": 8,
12
  "max_batch_size": 1,
13
  "device": "cuda",
14
- "cache_dir": "/tmp/models/tmpcdqz_pr1",
15
  "task": "",
16
  "save_load_fn": "hqq",
17
  "save_load_fn_args": {},
 
11
  "quant_hqq_weight_bits": 8,
12
  "max_batch_size": 1,
13
  "device": "cuda",
14
+ "cache_dir": "/tmp/models/tmpuszyy52l",
15
  "task": "",
16
  "save_load_fn": "hqq",
17
  "save_load_fn_args": {},