Spaces:
Sleeping
Sleeping
Mark Liu
commited on
Commit
·
920ccf6
1
Parent(s):
961aff6
Update app.py
Browse files
app.py
CHANGED
@@ -84,9 +84,8 @@ with gr.Blocks(css=css) as demo:
|
|
84 |
|
85 |
|
86 |
with gr.Column(elem_id="col-container"):
|
87 |
-
gr.Markdown("""##
|
88 |
-
Using the
|
89 |
-
Prompt templates from [awesome-chatgpt-prompts](https://github.com/f/awesome-chatgpt-prompts).<br>
|
90 |
Current limit is 3000 tokens per conversation.""",
|
91 |
elem_id="header")
|
92 |
|
@@ -106,9 +105,6 @@ with gr.Blocks(css=css) as demo:
|
|
106 |
temperature = gr.Slider(minimum=0, maximum=2.0, value=0.7, step=0.1, interactive=True, label="Temperature (higher = more creative/chaotic)")
|
107 |
max_tokens = gr.Slider(minimum=100, maximum=4096, value=1000, step=1, interactive=True, label="Max tokens per response")
|
108 |
|
109 |
-
gr.HTML('''<br><br><br><center><a href="https://huggingface.co/spaces/anzorq/chatgpt-demo?duplicate=true"><img src="https://bit.ly/3gLdBN6" alt="Duplicate Space"></a>You can duplicate this Space.<br>
|
110 |
-
Don't forget to set your own <a href="https://platform.openai.com/account/api-keys">OpenAI API Key</a> environment variable in Settings.<br>
|
111 |
-
<p><img src="https://visitor-badge.glitch.me/badge?page_id=anzorq.chatgpt_api_demo_hf" alt="visitors"></p></center>''')
|
112 |
|
113 |
btn_submit.click(submit_message, [user_token, input_message, prompt_template, temperature, max_tokens, state], [input_message, chatbot, total_tokens_str, state])
|
114 |
input_message.submit(submit_message, [user_token, input_message, prompt_template, temperature, max_tokens, state], [input_message, chatbot, total_tokens_str, state])
|
|
|
84 |
|
85 |
|
86 |
with gr.Column(elem_id="col-container"):
|
87 |
+
gr.Markdown("""## ChatGPT Online
|
88 |
+
Using the Official API (gpt-3.5-turbo model)<br>
|
|
|
89 |
Current limit is 3000 tokens per conversation.""",
|
90 |
elem_id="header")
|
91 |
|
|
|
105 |
temperature = gr.Slider(minimum=0, maximum=2.0, value=0.7, step=0.1, interactive=True, label="Temperature (higher = more creative/chaotic)")
|
106 |
max_tokens = gr.Slider(minimum=100, maximum=4096, value=1000, step=1, interactive=True, label="Max tokens per response")
|
107 |
|
|
|
|
|
|
|
108 |
|
109 |
btn_submit.click(submit_message, [user_token, input_message, prompt_template, temperature, max_tokens, state], [input_message, chatbot, total_tokens_str, state])
|
110 |
input_message.submit(submit_message, [user_token, input_message, prompt_template, temperature, max_tokens, state], [input_message, chatbot, total_tokens_str, state])
|