Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -30,6 +30,7 @@ MODEL_PATH = str(mistral_models_path)
|
|
30 |
tokenizer = MistralTokenizer.from_file(os.path.join(MODEL_PATH, "tokenizer.model.v3"))
|
31 |
model = Transformer.from_folder(MODEL_PATH)
|
32 |
|
|
|
33 |
@spaces.GPU()
|
34 |
def generate_response(message, history):
|
35 |
# Convert history to the format expected by the model
|
@@ -46,13 +47,14 @@ def generate_response(message, history):
|
|
46 |
tokens = tokenizer.encode_chat_completion(completion_request).tokens
|
47 |
|
48 |
# Generate response
|
49 |
-
out_tokens
|
50 |
|
51 |
# Decode response
|
52 |
result = tokenizer.instruct_tokenizer.tokenizer.decode(out_tokens[0])
|
53 |
|
54 |
return result
|
55 |
|
|
|
56 |
# Gradio interface
|
57 |
iface = gr.ChatInterface(
|
58 |
generate_response,
|
|
|
30 |
tokenizer = MistralTokenizer.from_file(os.path.join(MODEL_PATH, "tokenizer.model.v3"))
|
31 |
model = Transformer.from_folder(MODEL_PATH)
|
32 |
|
33 |
+
|
34 |
@spaces.GPU()
|
35 |
def generate_response(message, history):
|
36 |
# Convert history to the format expected by the model
|
|
|
47 |
tokens = tokenizer.encode_chat_completion(completion_request).tokens
|
48 |
|
49 |
# Generate response
|
50 |
+
out_tokens = generate([tokens], model, max_tokens=256, temperature=0.7, eos_id=tokenizer.instruct_tokenizer.tokenizer.eos_id)
|
51 |
|
52 |
# Decode response
|
53 |
result = tokenizer.instruct_tokenizer.tokenizer.decode(out_tokens[0])
|
54 |
|
55 |
return result
|
56 |
|
57 |
+
|
58 |
# Gradio interface
|
59 |
iface = gr.ChatInterface(
|
60 |
generate_response,
|