BhumikaMak commited on
Commit
2aea1e4
·
verified ·
1 Parent(s): 4dfed64

update: change bg colour, revert-2

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -162,10 +162,10 @@ body {
162
  with gr.Blocks(css=custom_css) as interface:
163
 
164
  gr.HTML("""
165
- <span style="color: #800000; font-family: 'Papyrus', cursive; font-weight: bold; font-size: 32px; background-color: #d3d3d3;">NeuralVista</span><br><br>
166
-
167
-
168
- <span style="color: black; font-family: 'Papyrus', cursive; font-size: 18px; background-color: #d3d3d3;">A harmonious framework of tools ☼ designed to illuminate the inner workings of AI.</span>
169
  """)
170
 
171
  # Default sample
@@ -201,7 +201,7 @@ with gr.Blocks(css=custom_css) as interface:
201
  )
202
 
203
 
204
- gr.HTML("""<span style="font-family: 'Papyrus', cursive; font-size: 14px; background-color: #d3d3d3;">The visualization demonstrates object detection and interpretability. Detected objects are highlighted with bounding boxes, while the heatmap reveals regions of focus, offering insights into the model's decision-making process.</span>""")
205
  # Results and visualization
206
  with gr.Row(elem_classes="custom-row"):
207
  result_gallery = gr.Gallery(
@@ -223,9 +223,9 @@ with gr.Blocks(css=custom_css) as interface:
223
 
224
  gr.HTML("""
225
  <span style="font-family: 'Papyrus', cursive; font-size: 14px;">
226
- <span style="color: #800000 background-color: #d3d3d3; ;">Concept Discovery</span> is the process of uncovering the hidden, high-level features that a deep learning model has learned. It provides a way to understand the essence of its internal representations, akin to peering into the mind of the model and revealing the meaningful patterns it detects in the data.
227
  <br><br>
228
- <span style="color: #800000 background-color: #d3d3d3;;">Deep Feature Factorization</span> (DFF) serves as a tool for breaking down these complex features into simpler, more interpretable components. By applying matrix factorization on activation maps, it untangles the intricate web of learned representations, making it easier to comprehend what the model is truly focusing on. Together, these methods bring us closer to understanding the underlying logic of neural networks, shedding light on the often enigmatic decisions they make.
229
  </span>
230
  """)
231
 
 
162
  with gr.Blocks(css=custom_css) as interface:
163
 
164
  gr.HTML("""
165
+ <div style="background-color: #d3d3d3; padding: 20px; border-radius: 8px;">
166
+ <span style="color: #800000; font-family: 'Papyrus', cursive; font-weight: bold; font-size: 32px;">NeuralVista</span><br><br>
167
+ <span style="color: black; font-family: 'Papyrus', cursive; font-size: 18px;">A harmonious framework of tools ☼ designed to illuminate the inner workings of AI.</span>
168
+ </div>
169
  """)
170
 
171
  # Default sample
 
201
  )
202
 
203
 
204
+ gr.HTML("""<span style="font-family: 'Papyrus', cursive; font-size: 14px;">The visualization demonstrates object detection and interpretability. Detected objects are highlighted with bounding boxes, while the heatmap reveals regions of focus, offering insights into the model's decision-making process.</span>""")
205
  # Results and visualization
206
  with gr.Row(elem_classes="custom-row"):
207
  result_gallery = gr.Gallery(
 
223
 
224
  gr.HTML("""
225
  <span style="font-family: 'Papyrus', cursive; font-size: 14px;">
226
+ <span style="color: #800000;">Concept Discovery</span> is the process of uncovering the hidden, high-level features that a deep learning model has learned. It provides a way to understand the essence of its internal representations, akin to peering into the mind of the model and revealing the meaningful patterns it detects in the data.
227
  <br><br>
228
+ <span style="color: #800000;">Deep Feature Factorization</span> (DFF) serves as a tool for breaking down these complex features into simpler, more interpretable components. By applying matrix factorization on activation maps, it untangles the intricate web of learned representations, making it easier to comprehend what the model is truly focusing on. Together, these methods bring us closer to understanding the underlying logic of neural networks, shedding light on the often enigmatic decisions they make.
229
  </span>
230
  """)
231