Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -2,12 +2,13 @@ import streamlit as st
|
|
2 |
from transformers import pipeline
|
3 |
st.header("Ways to Improve Your Conversational Agents using π€ Hugging Face")
|
4 |
|
5 |
-
st.write("There are many ways to improve your conversational agents using language models. In this blog post, I will walk you through a couple of tricks that will improve your conversational agent.")
|
|
|
6 |
|
7 |
st.subheader("Data Augmentation with Generative Models β¨")
|
8 |
st.write("There are cases where you will not be allowed to keep data, you will have to start from scratch or you will have very little amount of data. We'll go over two use cases and see how to tackle them.")
|
9 |
st.write("Imagine you're making a chatbot that will answer very general questions about emergency situations at home.")
|
10 |
-
st.write("If you have very little amount of data, you could actually augment it through language models. There are regex based tools you can use but they tend to create bias due to repetitive patterns, so it's better to use language models for this case. A good model to use is a generative model fine-tuned on Quora Question Pairs dataset. This dataset consists of question pairs that are paraphrase of one another, and T5 can generate a paraphrased question given a source question.")
|
11 |
st.write("Try it yourself here ππ»")
|
12 |
|
13 |
|
@@ -27,12 +28,25 @@ st.write("Your English intent classification model will be between these two mod
|
|
27 |
|
28 |
translator = pipeline("translation", model="Helsinki-NLP/opus-mt-en-fr")
|
29 |
default_value_tr = "How are you?"
|
30 |
-
input = st.text_input("Input", default_value_tr, key = "translation")
|
31 |
outputs = translator(input)
|
32 |
st.write("Translated Example:")
|
33 |
st.write(outputs[0]["translation_text"])
|
34 |
st.write("You can check out this [link](https://huggingface.co/models?pipeline_tag=translation&sort=downloads&search=helsinki-nlp) for available translation models.")
|
35 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
st.subheader("Add Personas to Your Conversational Agent using GPT-2")
|
37 |
-
st.write("When trained, language models like GPT-2 or DialoGPT is capable of talking like any character you want. If you have a friend-like chatbot (instead of a chatbot built for RPA) you can give your users options to talk to their favorite character. There are couple of ways of doing this, you can either fine-tune DialoGPT with sequences of conversation turns, maybe movie dialogues, or infer with a large model like GPT-J. ")
|
38 |
-
st.write("You can see an [example](https://huggingface.co/docs/transformers/model_doc/dialogpt) of a chatbot that talks like Gandalf, that is done simply by sending a request to GPT-J through Inference API.")
|
|
|
|
|
|
2 |
from transformers import pipeline
|
3 |
st.header("Ways to Improve Your Conversational Agents using π€ Hugging Face")
|
4 |
|
5 |
+
st.write("There are many ways to improve your conversational agents using language models. In this blog post, I will walk you through a couple of tricks that will improve your conversational agent. πΎ")
|
6 |
+
st.write("There are multiple ways of building conversational agents, you can build intent-action based chatbots for automating processes or build language-model based chatbots like [DialoGPT](https://huggingface.co/docs/transformers/model_doc/dialogpt). This blog post is for intent-action based agents, but in the last part, I'm talking about a hybrid agent that switches between our chatbot and a dialogue language model.")
|
7 |
|
8 |
st.subheader("Data Augmentation with Generative Models β¨")
|
9 |
st.write("There are cases where you will not be allowed to keep data, you will have to start from scratch or you will have very little amount of data. We'll go over two use cases and see how to tackle them.")
|
10 |
st.write("Imagine you're making a chatbot that will answer very general questions about emergency situations at home.")
|
11 |
+
st.write("If you have very little amount of data, you could actually augment it through language models. There are regex based tools you can use but they tend to create bias due to repetitive patterns, so it's better to use language models for this case. A good model to use is a generative model fine-tuned on [Quora Question Pairs dataset](https://www.tensorflow.org/datasets/catalog/glue?hl=en#glueqqp). This dataset consists of question pairs that are paraphrase of one another, and T5 can generate a paraphrased question given a source question. There's a similar dataset called [MRPC](https://www.tensorflow.org/datasets/catalog/glue?hl=en#gluemrpc) that assesses if one sentence is a paraphrase of another, you can choose between one of them.")
|
12 |
st.write("Try it yourself here ππ»")
|
13 |
|
14 |
|
|
|
28 |
|
29 |
translator = pipeline("translation", model="Helsinki-NLP/opus-mt-en-fr")
|
30 |
default_value_tr = "How are you?"
|
31 |
+
input = st.text_input("Input in English", default_value_tr, key = "translation")
|
32 |
outputs = translator(input)
|
33 |
st.write("Translated Example:")
|
34 |
st.write(outputs[0]["translation_text"])
|
35 |
st.write("You can check out this [link](https://huggingface.co/models?pipeline_tag=translation&sort=downloads&search=helsinki-nlp) for available translation models.")
|
36 |
|
37 |
+
|
38 |
+
st.subheader("Easy Information Retrieval")
|
39 |
+
st.write("If you're making a chatbot that needs to provide information to user, you can take user's query and search for the answer in the documents you have, using question answering models. Look at the example and try it yourself here ππ»")
|
40 |
+
|
41 |
+
qa_model = pipeline("question-answering")
|
42 |
+
question = st.text_input("Question", default_value = "What does transformers do?")
|
43 |
+
context = st.text_area("Context", default_value = "π€ Transformers provides thousands of pretrained models to perform tasks on different modalities such as text, vision, and audio.")
|
44 |
+
output_answer = qa_model(question = question, context = context)
|
45 |
+
st.write("Answer:")
|
46 |
+
st.write(output_answer["answer"])
|
47 |
+
|
48 |
st.subheader("Add Personas to Your Conversational Agent using GPT-2")
|
49 |
+
st.write("When trained, language models like GPT-2 or DialoGPT is capable of talking like any character you want. If you have a friend-like chatbot (instead of a chatbot built for RPA) you can give your users options to talk to their favorite character. There are couple of ways of doing this, you can either fine-tune DialoGPT with sequences of conversation turns, maybe movie dialogues, or infer with a large model like GPT-J. Note that these models might have biases and you will not have any control over output, unless you make an additional effort to filter it.")
|
50 |
+
st.write("You can see an [example](https://huggingface.co/docs/transformers/model_doc/dialogpt) of a chatbot that talks like Gandalf, that is done simply by sending a request to GPT-J through Inference API.")
|
51 |
+
|
52 |
+
I've written the inferences in this blog post with only three lines of code, using [pipelines](https://huggingface.co/docs/transformers/main_classes/pipelines). (yes π€―) Check out the code of the post [here](https://huggingface.co/spaces/merve/chatbot-blog/blob/main/app.py) on how you can do it too! π€
|