Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -49,7 +49,9 @@ def run_generation(user_text, use_assistant, temperature, max_new_tokens):
|
|
49 |
model_output = ""
|
50 |
for new_text in streamer:
|
51 |
model_output += new_text
|
52 |
-
|
|
|
|
|
53 |
|
54 |
|
55 |
def reset_textbox():
|
@@ -82,11 +84,11 @@ with gr.Blocks() as demo:
|
|
82 |
temperature = gr.Slider(
|
83 |
minimum=0.0, maximum=2.0, value=0.6, step=0.05, interactive=True, label="Temperature (0.0 = Greedy)",
|
84 |
)
|
85 |
-
gr.Markdown("###
|
86 |
-
|
87 |
|
88 |
generate_inputs = [user_text, use_assistant, temperature, max_new_tokens]
|
89 |
-
generate_outputs = [model_output,
|
90 |
user_text.submit(run_generation, generate_inputs, generate_outputs)
|
91 |
button_submit.click(run_generation, generate_inputs, generate_outputs)
|
92 |
|
|
|
49 |
model_output = ""
|
50 |
for new_text in streamer:
|
51 |
model_output += new_text
|
52 |
+
time_so_far = round(time.time() - start, 3)
|
53 |
+
tokens_so_far = tokenizer(model_output, return_tensors="pt").input_ids.shape[1]
|
54 |
+
yield [model_output, tokens_so_far/time_so_far]
|
55 |
|
56 |
|
57 |
def reset_textbox():
|
|
|
84 |
temperature = gr.Slider(
|
85 |
minimum=0.0, maximum=2.0, value=0.6, step=0.05, interactive=True, label="Temperature (0.0 = Greedy)",
|
86 |
)
|
87 |
+
gr.Markdown("### Tokens per secon")
|
88 |
+
tokens_per_second = gr.Textbox(lines=1, interactive=False, show_label=False)
|
89 |
|
90 |
generate_inputs = [user_text, use_assistant, temperature, max_new_tokens]
|
91 |
+
generate_outputs = [model_output, tokens_per_second]
|
92 |
user_text.submit(run_generation, generate_inputs, generate_outputs)
|
93 |
button_submit.click(run_generation, generate_inputs, generate_outputs)
|
94 |
|