Kkordik commited on
Commit
be7c477
·
verified ·
1 Parent(s): 9afd959

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -1
app.py CHANGED
@@ -30,6 +30,7 @@ MODEL_PATH = str(mistral_models_path)
30
  tokenizer = MistralTokenizer.from_file(os.path.join(MODEL_PATH, "tokenizer.model.v3"))
31
  model = Transformer.from_folder(MODEL_PATH)
32
 
 
33
  @spaces.GPU()
34
  def generate_response(message, history):
35
  # Convert history to the format expected by the model
@@ -46,13 +47,14 @@ def generate_response(message, history):
46
  tokens = tokenizer.encode_chat_completion(completion_request).tokens
47
 
48
  # Generate response
49
- out_tokens, * = generate([tokens], model, max_tokens=256, temperature=0.7, eos_id=tokenizer.instruct_tokenizer.tokenizer.eos_id)
50
 
51
  # Decode response
52
  result = tokenizer.instruct_tokenizer.tokenizer.decode(out_tokens[0])
53
 
54
  return result
55
 
 
56
  # Gradio interface
57
  iface = gr.ChatInterface(
58
  generate_response,
 
30
  tokenizer = MistralTokenizer.from_file(os.path.join(MODEL_PATH, "tokenizer.model.v3"))
31
  model = Transformer.from_folder(MODEL_PATH)
32
 
33
+
34
  @spaces.GPU()
35
  def generate_response(message, history):
36
  # Convert history to the format expected by the model
 
47
  tokens = tokenizer.encode_chat_completion(completion_request).tokens
48
 
49
  # Generate response
50
+ out_tokens = generate([tokens], model, max_tokens=256, temperature=0.7, eos_id=tokenizer.instruct_tokenizer.tokenizer.eos_id)
51
 
52
  # Decode response
53
  result = tokenizer.instruct_tokenizer.tokenizer.decode(out_tokens[0])
54
 
55
  return result
56
 
57
+
58
  # Gradio interface
59
  iface = gr.ChatInterface(
60
  generate_response,