Spaces:
Sleeping
Sleeping
File size: 1,391 Bytes
5eff8c0 f13ba47 5eff8c0 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 |
from huggingface_hub import InferenceClient
import gradio as gr
"""
For more information on `huggingface_hub` Inference API support, please check the docs: https://huggingface.co/docs/huggingface_hub/v0.22.2/en/guides/inference
"""
client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
def respond(
prompt: str,
history,
):
messages = [{"role": "system", "content": "You are a friendly chatbot"}]
for val in history:
if val[0]:
messages.append({"role": "user", "content": val[0]})
if val[1]:
messages.append({"role": "assistant", "content": val[1]})
yield messages
messages.append({"role": "user", "content": prompt})
history.append([prompt, None])
response = ""
for message in client.chat_completion(
messages,
stream=True,
):
response += message.choices[0].delta.content or ""
history[-1][1] = response
yield history
with gr.Blocks() as demo:
gr.Markdown("# Chat with Hugging Face Zephyr 7b 🤗")
chatbot = gr.Chatbot(
label="Agent",
avatar_images=(
None,
"https://em-content.zobj.net/source/twitter/376/hugging-face_1f917.png",
),
)
prompt = gr.Textbox(lines=1, label="Chat Message")
prompt.submit(respond, [prompt, chatbot], [chatbot])
if __name__ == "__main__":
demo.launch()
|