ashok2216's picture
Update app.py
7713456 verified
raw
history blame
4.56 kB
import pybase64 as base64
import pandas as pd
import streamlit as st
import seaborn as sns
from data_cleaning import preprocess
from transformers import pipeline
from data_integration import scrape_all_pages
#@st.cache_data
#def get_img_as_base64(file):
# with open(file, "rb") as f:
# data = f.read()
# return base64.b64encode(data).decode()
#img = get_img_as_base64("image.jpg")background-image: url("data:image/png;base64,{img}");
page_bg_img = f"""
<style>
[data-testid="stAppViewContainer"] > .main {{
background-image: url("data:image/jpeg;base64,/9j/4AAQSkZJRgABAQAAAQABAAD/2wBDAAsJCQcJCQcJCQkJCwkJCQkJCQsJCwsMCwsLDA0QDBEODQ4MEhkSJRodJR0ZHxwpKRYlNzU2GioyPi0pMBk7IRP/2wBDAQcICAsJCxULCxUsHRkdLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCwsLCz/wAARCAC0AQMDASIAAhEBAxEB/8QAGgAAAwEBAQEAAAAAAAAAAAAAAAECAwQFBv/EACIQAQEBAQACAwEBAAMBAAAAAAABEQJRYRIhMUFxgZGhwf/EABoBAAMBAQEBAAAAAAAAAAAAAAECAwAFBAb/xAAYEQEBAQEBAAAAAAAAAAAAAAAAARECEv/aAAwDAQACEQMRAD8A9aW+VS3ymKj6p4Ke3yc3yRwZUqf35Pb5AwUqNvk9vkYchpUaUt8nLfIkPDanRt8nN8jDkbSUbfKtvksVIGlG3yNvkYeNrYNo2gmHD0aQYcPaqVBlp8aS36+1S3yzi5S0ZGvNvmtObfLGVpKh1DRvzb5aS3yw5ac1CwzeXryuW+WMq5UbGa719/d/7Pb5qNOUmCrb5CdAYz5FUhKjq691EVIUVP4Op08PDhzG1KlIeHh5R1KwsOQ8PKOp2FgxWU8bS2JkPDw8raXCAwDo4VI6Q6aQAgJpDCdP/kDeVz9VP6zipS1sbStJWMrSVLocbc1pyx5rSfxGwGsq5WU/FypWNrSX9PWct+zlpLA1e0J0Bga+W1cZyrl/+Og6VXF5PpEVA0lUqRMXPAalYchyCKg6nRIeHDkjaSwsGKyDI2lvKchqyDI2hiLCXf8AU3/RlHynE1abDynkTSOkOmnIBAdP5VFS1EVKWh5aStJWUq5U6GNea0l/GMXKnSVtKuXGMvtU69ksJWsv6es9/fsfL2XCNNCN9hsB81KvmspWkr1uvWsv4uMpfxpKSp2NIuIi4BKuKiYqBqdiocKKhdDADNtLiQrBja3lBVWFTaOIqaupppTzlCaqlTynnKQYbTYDIw0PK5VREXC0li5Vys5VShUrGspys9h6VOtNGs9GtidafIM9gbAx89K05rCVpzV67Dfm/jTmsOa05paWxvKvllzWkvslJY1i4yl9rhC40ikSrLpcOKTFF1sBGG1vKaVUmjppyipXU00p5yipq6k0qk5SDBvQ+SOANoXlUqomKjalYqfxUQcBLqK09Ro1ojYvS+SNK08SsafIM9oHCZXgSrlY832uX2o7Vjo5rTmufmteb7JSt+a1lc/NayksLY2laSsZWkvtOhjaVUrKLl9koY100aqFHyYAKPkFRobTTlNRV1N/oyqTlNxNURp0fykHhD6byAPs4OhYIoAdSvIPSDajeT0rSKniN5FpWlb+ptUiV5P5BnvsHDy8PmtJWPNaS/RnWxrzWsrCVpzS0uN+a0lYStJSUMdErTmsOa0lSoY2laSsZVyp1sbSq1nKqUlppyv7NMpktN5ABF085FTTpNqk5KkdLPZpTeSGGB1vJGDHS2EeDDHUuoRHSptR65JFqrWdqkqN5K32m0W/aOqryneRoRv+hQPLxouIi4d0cXy0jOLhKGNZWnNYytJpa2Nua05rGVpzUqGNua0lY81pKnRxtKqVlL7XKlTyNJT1Mpp1ScqBaCqTkUgGUnJWAyjaPkjw/oNoXkAHnodJYQMh1LqFf6m/iqmqSo3lNrPpdZ1WVKxN/rO1d/rLqrcp+S0J0KB5eZJioeHh3rOKhT+KkLQONJUSLhKzSNJWUaROi1lXKyjWJ00jSLlZxcRp5FyqlRFJ1aQ9NJlqshghhdPg/TH0A0cABjoWAwMGVKwEZGlSsTf1F/q6mqSpWM6irrPpWJWI6rK1fX9Z9LckvKQQU0McvxGLwYbVUyYo8PAtYSKkEVhAOLiZGkLRVGkZxpE6dcVCkVEapFRScUnVoDI/CdWkMAFUBkbCDGGxaAeDIMqVIqpNPEqmoq6zv9UiVR0y6a1l0rynYy6RV9M6tyWxILQoGJwfFeDA1kfE8VIeemBMVIeHIAiRchSKkLWVFxMi5E6pFRUJUSqnKp+GUNKq8mcIyLwGX2ZKoDKRTCDhRTEoBj+ClSv4irv4nDxKoqL/AFpWdn6pE6zrLpr0y6W5JWXTOtOmdW5KgAKM0+MHxgAEOcz2PjAGZU5mCcwAGXOZqpzACVl88z2ucz7ATPFfGf8Ai/jyAnVIqczP6PjAE6vyc5h/GQAlWgnM9nk+wE6pBOYr4z2AzU/jDnM9gMSneZ7HxgAxOj4xF5gBonUXmIsgCsSrLqRl1IAtyWsupGdkAX5IjIAFGf/Z");
background-size: 100%;
background-position: top left;
background-repeat: no-repeat;
background-attachment: local;
}}
[data-testid="stSidebar"] > div:first-child {{
background-position: center;
background-repeat: no-repeat;
background-attachment: fixed;
}}
[data-testid="stHeader"] {{
background: rgba(0,0,0,0);
}}
[data-testid="stToolbar"] {{
right: 2rem;
}}
</style>
"""
st.markdown(page_bg_img, unsafe_allow_html=True)
# st.image("logo.png", width=200, height=200)
st.image("logo.png", width=100)
st.subheader(':blue[NLP HUB®]')
st.header('Amazon Sentiment Analysis using FineTuned :green[GPT-2] Pre-Trained Model')
sentiment_model = pipeline(model="ashok2216/gpt2-amazon-sentiment-classifier")
# Example usage:-
sample_url = 'https://www.amazon.in/Dell-Inspiron-i7-1255U-Processor-Platinum/product-reviews/B0C9F142V6/ref=cm_cr_dp_d_show_all_btm?ie=UTF8&reviewerType=all_reviews'
url = st.text_input("Amazon product link", sample_url)
st.write("Done")
st.subheader('', divider='rainbow')
all_reviews = scrape_all_pages(url)
# Convert to DataFrame for further analysis
reviews = pd.DataFrame(all_reviews)
reviews['processed_text'] = reviews['content'].apply(preprocess)
# st.dataframe(reviews, use_container_width=True)
# st.markdown(sentiment_model(['It is Super!']))
sentiments = []
for text in reviews['processed_text']:
if list(sentiment_model(text)[0].values())[0] == 'LABEL_1':
output = 'Positive'
else:
output = 'Negative'
sentiments.append(output)
reviews['sentiments'] = sentiments
st.markdown(':rainbow[Output]')
st.dataframe(reviews, use_container_width=True)
# sns.countplot(reviews['sentiments'])