Spaces:
Sleeping
Sleeping
Update webui.py
Browse files
webui.py
CHANGED
@@ -53,7 +53,6 @@ def load_model(path, n_ctx, n_gpu_layers, n_threads, verbose, f16_kv, logits_all
|
|
53 |
except:
|
54 |
return ""
|
55 |
|
56 |
-
load_model("llama-2-7b-chat.ggmlv3.q2_K.bin", 2048, 0, 32, True, True, False, False, True, False, 512, 64, False, 10000, 1)
|
57 |
|
58 |
def list_models(name):
|
59 |
return os.listdir(f"{dir}/models")
|
@@ -80,6 +79,8 @@ If a question does not make any sense, or is not factually coherent, explain why
|
|
80 |
"""
|
81 |
'''
|
82 |
|
|
|
|
|
83 |
#@spaces.GPU
|
84 |
def generate_text(message, history, system_prompt, preset, temperature, max_tokens, top_p, top_k, repeat_penalty):
|
85 |
temp = ""
|
@@ -115,7 +116,7 @@ def generate_text(message, history, system_prompt, preset, temperature, max_toke
|
|
115 |
|
116 |
|
117 |
|
118 |
-
chatbot = gr.Chatbot(show_label=False, layout="panel", show_copy_button=True, height=500, min_width=180)
|
119 |
|
120 |
with gr.Blocks(theme="theme-repo/STONE_Theme", title="TensorLM", css="style.css") as demo:
|
121 |
with gr.Row():
|
|
|
53 |
except:
|
54 |
return ""
|
55 |
|
|
|
56 |
|
57 |
def list_models(name):
|
58 |
return os.listdir(f"{dir}/models")
|
|
|
79 |
"""
|
80 |
'''
|
81 |
|
82 |
+
load_model("llama-2-7b-chat.ggmlv3.q2_K.bin", 2048, 0, 32, True, True, False, False, True, False, 512, 64, False, 10000, 1)
|
83 |
+
|
84 |
#@spaces.GPU
|
85 |
def generate_text(message, history, system_prompt, preset, temperature, max_tokens, top_p, top_k, repeat_penalty):
|
86 |
temp = ""
|
|
|
116 |
|
117 |
|
118 |
|
119 |
+
chatbot = gr.Chatbot(show_label=False, layout="panel", show_copy_button=True, height=500, min_width=180, autofocus=False)
|
120 |
|
121 |
with gr.Blocks(theme="theme-repo/STONE_Theme", title="TensorLM", css="style.css") as demo:
|
122 |
with gr.Row():
|