Fix UI issues and improve user experience
Browse filesBug fixes:
- Fix collapsible code blocks using native Streamlit expander
- Fix image display to show actual images instead of filenames
- Remove persistent "Data loaded successfully!" message
UI improvements:
- Reduce excessive whitespace and margins for tighter layout
- Move Quick Queries to top of sidebar for better accessibility
- Compact message styling with reduced padding
- Cleaner dataset info section with condensed layout
π€ Generated with [Claude Code](https://claude.ai/code)
Co-Authored-By: Claude <[email protected]>
app.py
CHANGED
@@ -106,34 +106,24 @@ st.markdown("""
|
|
106 |
.user-message {
|
107 |
background: #3b82f6;
|
108 |
color: white;
|
109 |
-
padding: 1rem
|
110 |
border-radius: 12px;
|
111 |
-
margin: 2rem 0;
|
112 |
-
margin-left: auto;
|
113 |
-
margin-right: 0;
|
114 |
max-width: 70%;
|
115 |
-
display: flex;
|
116 |
-
justify-content: flex-end;
|
117 |
}
|
118 |
|
119 |
.user-info {
|
120 |
font-size: 0.875rem;
|
121 |
opacity: 0.9;
|
122 |
-
margin-bottom:
|
123 |
}
|
124 |
|
125 |
/* Assistant message styling */
|
126 |
.assistant-message {
|
127 |
background: #f1f5f9;
|
128 |
color: #334155;
|
129 |
-
padding: 1rem
|
130 |
border-radius: 12px;
|
131 |
-
margin: 2rem 0;
|
132 |
-
margin-left: 0;
|
133 |
-
margin-right: auto;
|
134 |
max-width: 70%;
|
135 |
-
display: flex;
|
136 |
-
justify-content: flex-start;
|
137 |
}
|
138 |
|
139 |
.assistant-info {
|
@@ -503,7 +493,7 @@ st.markdown("<hr style='margin: 1rem 0; border: none; border-top: 1px solid #e2e
|
|
503 |
# Load data with error handling
|
504 |
try:
|
505 |
df = preprocess_and_load_df(join(self_path, "Data.csv"))
|
506 |
-
|
507 |
except Exception as e:
|
508 |
st.error(f"Error loading data: {e}")
|
509 |
st.stop()
|
@@ -511,43 +501,9 @@ except Exception as e:
|
|
511 |
inference_server = "https://api-inference.huggingface.co/models/mistralai/Mistral-7B-Instruct-v0.2"
|
512 |
image_path = "IITGN_Logo.png"
|
513 |
|
514 |
-
# Clean sidebar
|
515 |
with st.sidebar:
|
516 |
-
#
|
517 |
-
st.markdown("### Dataset Info")
|
518 |
-
st.markdown("""
|
519 |
-
<div style='background-color: #f1f5f9; padding: 1rem; border-radius: 8px; margin-bottom: 1.5rem;'>
|
520 |
-
<h4 style='margin: 0 0 0.5rem 0; color: #1e293b; font-size: 1rem;'>PM2.5 Air Quality Data</h4>
|
521 |
-
<p style='margin: 0.25rem 0; font-size: 0.875rem;'><strong>Time Range:</strong> Daily measurements</p>
|
522 |
-
<p style='margin: 0.25rem 0; font-size: 0.875rem;'><strong>Locations:</strong> Multiple cities in Gujarat</p>
|
523 |
-
<p style='margin: 0.25rem 0; font-size: 0.875rem;'><strong>Records:</strong> Air quality monitoring data</p>
|
524 |
-
<p style='margin: 0.25rem 0; font-size: 0.875rem;'><strong>Parameters:</strong> PM2.5, PM10, Location data</p>
|
525 |
-
</div>
|
526 |
-
""", unsafe_allow_html=True)
|
527 |
-
|
528 |
-
# Current Model Info
|
529 |
-
st.markdown("### Current Model")
|
530 |
-
st.markdown(f"**{model_name}**")
|
531 |
-
|
532 |
-
model_descriptions = {
|
533 |
-
"llama3.1": "Fast and efficient for general queries",
|
534 |
-
"llama3.3": "Most advanced LLaMA model for complex reasoning",
|
535 |
-
"mistral": "Balanced performance and speed",
|
536 |
-
"gemma": "Google's lightweight model",
|
537 |
-
"gemini-pro": "Google's most powerful model",
|
538 |
-
"gpt-oss-20b": "OpenAI's compact open-weight GPT for everyday tasks",
|
539 |
-
"gpt-oss-120b": "OpenAI's massive open-weight GPT for nuanced responses",
|
540 |
-
"deepseek-R1": "DeepSeek's distilled LLaMA model for efficient reasoning",
|
541 |
-
"llama4 maverik": "Meta's LLaMA 4 Maverick β high-performance instruction model",
|
542 |
-
"llama4 scout": "Meta's LLaMA 4 Scout β optimized for adaptive reasoning"
|
543 |
-
}
|
544 |
-
|
545 |
-
if model_name in model_descriptions:
|
546 |
-
st.caption(model_descriptions[model_name])
|
547 |
-
|
548 |
-
st.markdown("---")
|
549 |
-
|
550 |
-
# Quick Queries Section
|
551 |
st.markdown("### Quick Queries")
|
552 |
|
553 |
# Load quick prompts
|
@@ -585,6 +541,38 @@ with st.sidebar:
|
|
585 |
|
586 |
st.markdown("---")
|
587 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
588 |
# Clear Chat Button
|
589 |
if st.button("Clear Chat", use_container_width=True):
|
590 |
st.session_state.responses = []
|
@@ -609,18 +597,18 @@ def show_custom_response(response):
|
|
609 |
content = response.get("content", "")
|
610 |
|
611 |
if role == "user":
|
612 |
-
# User message with right alignment
|
613 |
st.markdown(f"""
|
614 |
-
<div style='display: flex; justify-content: flex-end; margin:
|
615 |
<div class='user-message'>
|
616 |
{content}
|
617 |
</div>
|
618 |
</div>
|
619 |
""", unsafe_allow_html=True)
|
620 |
elif role == "assistant":
|
621 |
-
# Assistant message with left alignment
|
622 |
st.markdown(f"""
|
623 |
-
<div style='display: flex; justify-content: flex-start; margin:
|
624 |
<div class='assistant-message'>
|
625 |
<div class='assistant-info'>VayuChat</div>
|
626 |
{content if isinstance(content, str) else str(content)}
|
@@ -628,32 +616,28 @@ def show_custom_response(response):
|
|
628 |
</div>
|
629 |
""", unsafe_allow_html=True)
|
630 |
|
631 |
-
# Show generated code with
|
632 |
if response.get("gen_code"):
|
633 |
-
st.
|
634 |
-
|
635 |
-
<div class='code-header' onclick='toggleCode(this)'>
|
636 |
-
<div class='code-title'>Generated Python Code</div>
|
637 |
-
<div class='toggle-text'>Click to expand</div>
|
638 |
-
</div>
|
639 |
-
<div class='code-block' style='display: none;'>
|
640 |
-
""", unsafe_allow_html=True)
|
641 |
-
|
642 |
-
st.code(response["gen_code"], language="python")
|
643 |
-
|
644 |
-
st.markdown("</div></div>", unsafe_allow_html=True)
|
645 |
|
646 |
# Try to display image if content is a file path
|
647 |
try:
|
648 |
if isinstance(content, str) and (content.endswith('.png') or content.endswith('.jpg')):
|
649 |
if os.path.exists(content):
|
650 |
-
#
|
651 |
-
st.
|
652 |
-
<div style='background: white; border: 1px solid #e2e8f0; border-radius: 8px; padding: 1.5rem; margin: 1rem 0;'>
|
653 |
-
""", unsafe_allow_html=True)
|
654 |
-
st.image(content)
|
655 |
-
st.markdown("</div>", unsafe_allow_html=True)
|
656 |
return {"is_image": True}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
657 |
except:
|
658 |
pass
|
659 |
|
|
|
106 |
.user-message {
|
107 |
background: #3b82f6;
|
108 |
color: white;
|
109 |
+
padding: 0.75rem 1rem;
|
110 |
border-radius: 12px;
|
|
|
|
|
|
|
111 |
max-width: 70%;
|
|
|
|
|
112 |
}
|
113 |
|
114 |
.user-info {
|
115 |
font-size: 0.875rem;
|
116 |
opacity: 0.9;
|
117 |
+
margin-bottom: 3px;
|
118 |
}
|
119 |
|
120 |
/* Assistant message styling */
|
121 |
.assistant-message {
|
122 |
background: #f1f5f9;
|
123 |
color: #334155;
|
124 |
+
padding: 0.75rem 1rem;
|
125 |
border-radius: 12px;
|
|
|
|
|
|
|
126 |
max-width: 70%;
|
|
|
|
|
127 |
}
|
128 |
|
129 |
.assistant-info {
|
|
|
493 |
# Load data with error handling
|
494 |
try:
|
495 |
df = preprocess_and_load_df(join(self_path, "Data.csv"))
|
496 |
+
# Data loaded silently - no success message needed
|
497 |
except Exception as e:
|
498 |
st.error(f"Error loading data: {e}")
|
499 |
st.stop()
|
|
|
501 |
inference_server = "https://api-inference.huggingface.co/models/mistralai/Mistral-7B-Instruct-v0.2"
|
502 |
image_path = "IITGN_Logo.png"
|
503 |
|
504 |
+
# Clean sidebar
|
505 |
with st.sidebar:
|
506 |
+
# Quick Queries Section - moved to top
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
507 |
st.markdown("### Quick Queries")
|
508 |
|
509 |
# Load quick prompts
|
|
|
541 |
|
542 |
st.markdown("---")
|
543 |
|
544 |
+
# Dataset Info Section
|
545 |
+
st.markdown("### Dataset Info")
|
546 |
+
st.markdown("""
|
547 |
+
<div style='background-color: #f1f5f9; padding: 0.75rem; border-radius: 6px; margin-bottom: 1rem;'>
|
548 |
+
<h4 style='margin: 0 0 0.25rem 0; color: #1e293b; font-size: 0.9rem;'>PM2.5 Air Quality Data</h4>
|
549 |
+
<p style='margin: 0.125rem 0; font-size: 0.75rem;'><strong>Locations:</strong> Gujarat cities</p>
|
550 |
+
<p style='margin: 0.125rem 0; font-size: 0.75rem;'><strong>Parameters:</strong> PM2.5, PM10</p>
|
551 |
+
</div>
|
552 |
+
""", unsafe_allow_html=True)
|
553 |
+
|
554 |
+
# Current Model Info
|
555 |
+
st.markdown("### Current Model")
|
556 |
+
st.markdown(f"**{model_name}**")
|
557 |
+
|
558 |
+
model_descriptions = {
|
559 |
+
"llama3.1": "Fast and efficient for general queries",
|
560 |
+
"llama3.3": "Most advanced LLaMA model for complex reasoning",
|
561 |
+
"mistral": "Balanced performance and speed",
|
562 |
+
"gemma": "Google's lightweight model",
|
563 |
+
"gemini-pro": "Google's most powerful model",
|
564 |
+
"gpt-oss-20b": "OpenAI's compact open-weight GPT for everyday tasks",
|
565 |
+
"gpt-oss-120b": "OpenAI's massive open-weight GPT for nuanced responses",
|
566 |
+
"deepseek-R1": "DeepSeek's distilled LLaMA model for efficient reasoning",
|
567 |
+
"llama4 maverik": "Meta's LLaMA 4 Maverick β high-performance instruction model",
|
568 |
+
"llama4 scout": "Meta's LLaMA 4 Scout β optimized for adaptive reasoning"
|
569 |
+
}
|
570 |
+
|
571 |
+
if model_name in model_descriptions:
|
572 |
+
st.caption(model_descriptions[model_name])
|
573 |
+
|
574 |
+
st.markdown("---")
|
575 |
+
|
576 |
# Clear Chat Button
|
577 |
if st.button("Clear Chat", use_container_width=True):
|
578 |
st.session_state.responses = []
|
|
|
597 |
content = response.get("content", "")
|
598 |
|
599 |
if role == "user":
|
600 |
+
# User message with right alignment - reduced margins
|
601 |
st.markdown(f"""
|
602 |
+
<div style='display: flex; justify-content: flex-end; margin: 1rem 0;'>
|
603 |
<div class='user-message'>
|
604 |
{content}
|
605 |
</div>
|
606 |
</div>
|
607 |
""", unsafe_allow_html=True)
|
608 |
elif role == "assistant":
|
609 |
+
# Assistant message with left alignment - reduced margins
|
610 |
st.markdown(f"""
|
611 |
+
<div style='display: flex; justify-content: flex-start; margin: 1rem 0;'>
|
612 |
<div class='assistant-message'>
|
613 |
<div class='assistant-info'>VayuChat</div>
|
614 |
{content if isinstance(content, str) else str(content)}
|
|
|
616 |
</div>
|
617 |
""", unsafe_allow_html=True)
|
618 |
|
619 |
+
# Show generated code with Streamlit expander
|
620 |
if response.get("gen_code"):
|
621 |
+
with st.expander("π View Generated Code", expanded=False):
|
622 |
+
st.code(response["gen_code"], language="python")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
623 |
|
624 |
# Try to display image if content is a file path
|
625 |
try:
|
626 |
if isinstance(content, str) and (content.endswith('.png') or content.endswith('.jpg')):
|
627 |
if os.path.exists(content):
|
628 |
+
# Display image without showing filename
|
629 |
+
st.image(content, use_column_width=True)
|
|
|
|
|
|
|
|
|
630 |
return {"is_image": True}
|
631 |
+
# Also handle case where content shows filename but we want to show image
|
632 |
+
elif isinstance(content, str) and any(ext in content for ext in ['.png', '.jpg']):
|
633 |
+
# Extract potential filename from content
|
634 |
+
import re
|
635 |
+
filename_match = re.search(r'([^/\\]+\.(?:png|jpg|jpeg))', content)
|
636 |
+
if filename_match:
|
637 |
+
filename = filename_match.group(1)
|
638 |
+
if os.path.exists(filename):
|
639 |
+
st.image(filename, use_column_width=True)
|
640 |
+
return {"is_image": True}
|
641 |
except:
|
642 |
pass
|
643 |
|