wenhuach commited on
Commit
1d865bb
·
1 Parent(s): 90bb8ef

autoround format

Browse files

Signed-off-by: wenhuach <[email protected]>

config.json CHANGED
@@ -39,11 +39,11 @@
39
  "quantization_config": {
40
  "amp": true,
41
  "autoround_version": "0.4.2.dev",
 
42
  "batch_size": 8,
43
  "bits": 4,
44
- "damp_percent": 0.01,
45
  "data_type": "int",
46
- "desc_act": false,
47
  "enable_minmax_tuning": true,
48
  "enable_norm_bias_tuning": false,
49
  "enable_quanted_input": true,
@@ -54,7 +54,7 @@
54
  "lr": 0.001,
55
  "minmax_lr": 0.001,
56
  "nsamples": 512,
57
- "quant_method": "gptq",
58
  "scale_dtype": "torch.float16",
59
  "seqlen": 2048,
60
  "sym": true,
@@ -88,8 +88,7 @@
88
  "model.layers.25",
89
  "model.layers.26"
90
  ]
91
- ],
92
- "true_sequential": false
93
  },
94
  "rms_norm_eps": 1e-06,
95
  "rope_scaling": {
 
39
  "quantization_config": {
40
  "amp": true,
41
  "autoround_version": "0.4.2.dev",
42
+ "backend": "auto_round:gptq:exllamav2",
43
  "batch_size": 8,
44
  "bits": 4,
 
45
  "data_type": "int",
46
+ "dataset": "NeelNanda/pile-10k",
47
  "enable_minmax_tuning": true,
48
  "enable_norm_bias_tuning": false,
49
  "enable_quanted_input": true,
 
54
  "lr": 0.001,
55
  "minmax_lr": 0.001,
56
  "nsamples": 512,
57
+ "quant_method": "intel/auto-round",
58
  "scale_dtype": "torch.float16",
59
  "seqlen": 2048,
60
  "sym": true,
 
88
  "model.layers.25",
89
  "model.layers.26"
90
  ]
91
+ ]
 
92
  },
93
  "rms_norm_eps": 1e-06,
94
  "rope_scaling": {
quantize_config.json → quantization_config.json RENAMED
@@ -47,9 +47,8 @@
47
  ]
48
  ],
49
  "enable_norm_bias_tuning": false,
 
50
  "autoround_version": "0.4.2.dev",
51
- "quant_method": "gptq",
52
- "desc_act": false,
53
- "true_sequential": false,
54
- "damp_percent": 0.01
55
  }
 
47
  ]
48
  ],
49
  "enable_norm_bias_tuning": false,
50
+ "dataset": "NeelNanda/pile-10k",
51
  "autoround_version": "0.4.2.dev",
52
+ "quant_method": "intel/auto-round",
53
+ "backend": "auto_round:gptq:exllamav2"
 
 
54
  }