OPEA
/

GGUF
Inference Endpoints
conversational
cicdatopea commited on
Commit
7b32e98
·
verified ·
1 Parent(s): a85f4d5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -135,7 +135,7 @@ for n, m in block.named_modules():
135
 
136
  from auto_round import AutoRound
137
 
138
- autoround = AutoRound(model=model, tokenizer=tokenizer, layer_config=layer_config, device_map=device_map,
139
  iters=200,batch_size=8, seqlen=512)
140
  autoround.quantize()
141
  autoround.save_quantized(format="gguf:q4_0", output_dir="tmp_autoround"
 
135
 
136
  from auto_round import AutoRound
137
 
138
+ autoround = AutoRound(model=model, tokenizer=tokenizer, device_map=device_map,
139
  iters=200,batch_size=8, seqlen=512)
140
  autoround.quantize()
141
  autoround.save_quantized(format="gguf:q4_0", output_dir="tmp_autoround"