Tonic commited on
Commit
e63384d
·
1 Parent(s): d53c674

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -17,7 +17,7 @@ os.environ['PYTORCH_CUDA_ALLOC_CONF'] = 'max_split_size_mb:50'
17
  model_id = "TheBloke/Yi-34B-200K-Llamafied-GPTQ"
18
 
19
  gptq_config = GPTQConfig(bits=4, exllama_config={"version": 2})
20
- tokenizer = YiTokenizer.from_pretrained("./") #self-tokenizer method
21
  model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype="auto", trust_remote_code=True, quantization_config=gptq_config)
22
 
23
  def predict(message, max_new_tokens=4056, temperature=3.5, top_p=0.9, top_k=800):
 
17
  model_id = "TheBloke/Yi-34B-200K-Llamafied-GPTQ"
18
 
19
  gptq_config = GPTQConfig(bits=4, exllama_config={"version": 2})
20
+ tokenizer = YiTokenizer.from_pretrained("./")
21
  model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", torch_dtype="auto", trust_remote_code=True, quantization_config=gptq_config)
22
 
23
  def predict(message, max_new_tokens=4056, temperature=3.5, top_p=0.9, top_k=800):