David Driscoll
commited on
Commit
·
e421b40
1
Parent(s):
565e309
Update app
Browse files
app.py
CHANGED
|
@@ -131,7 +131,7 @@ custom_css = """
|
|
| 131 |
@import url('https://fonts.googleapis.com/css2?family=Orbitron:wght@400;700&display=swap');
|
| 132 |
body {
|
| 133 |
background-color: #0e0e0e;
|
| 134 |
-
color: #
|
| 135 |
font-family: 'Orbitron', sans-serif;
|
| 136 |
margin: 0;
|
| 137 |
padding: 0;
|
|
@@ -145,7 +145,7 @@ body {
|
|
| 145 |
}
|
| 146 |
.gradio-title {
|
| 147 |
font-size: 2.5em;
|
| 148 |
-
color: #
|
| 149 |
text-align: center;
|
| 150 |
margin-bottom: 0.2em;
|
| 151 |
}
|
|
@@ -153,6 +153,7 @@ body {
|
|
| 153 |
font-size: 1.2em;
|
| 154 |
text-align: center;
|
| 155 |
margin-bottom: 1em;
|
|
|
|
| 156 |
}
|
| 157 |
"""
|
| 158 |
|
|
@@ -164,7 +165,8 @@ posture_interface = gr.Interface(
|
|
| 164 |
inputs=gr.Image(sources=["webcam"], streaming=True, label="Capture Your Posture"),
|
| 165 |
outputs=[gr.Image(type="numpy", label="Annotated Output"), gr.Textbox(label="Posture Analysis")],
|
| 166 |
title="Posture Analysis",
|
| 167 |
-
description="Detects your posture using MediaPipe."
|
|
|
|
| 168 |
)
|
| 169 |
|
| 170 |
emotion_interface = gr.Interface(
|
|
@@ -172,7 +174,8 @@ emotion_interface = gr.Interface(
|
|
| 172 |
inputs=gr.Image(sources=["webcam"], streaming=True, label="Capture Your Face"),
|
| 173 |
outputs=[gr.Image(type="numpy", label="Annotated Output"), gr.Textbox(label="Emotion Analysis")],
|
| 174 |
title="Emotion Analysis",
|
| 175 |
-
description="Detects facial emotions using FER."
|
|
|
|
| 176 |
)
|
| 177 |
|
| 178 |
objects_interface = gr.Interface(
|
|
@@ -180,7 +183,8 @@ objects_interface = gr.Interface(
|
|
| 180 |
inputs=gr.Image(sources=["webcam"], streaming=True, label="Capture the Scene"),
|
| 181 |
outputs=[gr.Image(type="numpy", label="Annotated Output"), gr.Textbox(label="Object Detection")],
|
| 182 |
title="Object Detection",
|
| 183 |
-
description="Detects objects using a pretrained Faster R-CNN."
|
|
|
|
| 184 |
)
|
| 185 |
|
| 186 |
faces_interface = gr.Interface(
|
|
@@ -188,7 +192,8 @@ faces_interface = gr.Interface(
|
|
| 188 |
inputs=gr.Image(sources=["webcam"], streaming=True, label="Capture Your Face"),
|
| 189 |
outputs=[gr.Image(type="numpy", label="Annotated Output"), gr.Textbox(label="Face Detection")],
|
| 190 |
title="Face Detection",
|
| 191 |
-
description="Detects faces using MediaPipe."
|
|
|
|
| 192 |
)
|
| 193 |
|
| 194 |
# -----------------------------
|
|
|
|
| 131 |
@import url('https://fonts.googleapis.com/css2?family=Orbitron:wght@400;700&display=swap');
|
| 132 |
body {
|
| 133 |
background-color: #0e0e0e;
|
| 134 |
+
color: #ffffff;
|
| 135 |
font-family: 'Orbitron', sans-serif;
|
| 136 |
margin: 0;
|
| 137 |
padding: 0;
|
|
|
|
| 145 |
}
|
| 146 |
.gradio-title {
|
| 147 |
font-size: 2.5em;
|
| 148 |
+
color: #ffffff;
|
| 149 |
text-align: center;
|
| 150 |
margin-bottom: 0.2em;
|
| 151 |
}
|
|
|
|
| 153 |
font-size: 1.2em;
|
| 154 |
text-align: center;
|
| 155 |
margin-bottom: 1em;
|
| 156 |
+
color: #ffffff;
|
| 157 |
}
|
| 158 |
"""
|
| 159 |
|
|
|
|
| 165 |
inputs=gr.Image(sources=["webcam"], streaming=True, label="Capture Your Posture"),
|
| 166 |
outputs=[gr.Image(type="numpy", label="Annotated Output"), gr.Textbox(label="Posture Analysis")],
|
| 167 |
title="Posture Analysis",
|
| 168 |
+
description="Detects your posture using MediaPipe.",
|
| 169 |
+
live=True
|
| 170 |
)
|
| 171 |
|
| 172 |
emotion_interface = gr.Interface(
|
|
|
|
| 174 |
inputs=gr.Image(sources=["webcam"], streaming=True, label="Capture Your Face"),
|
| 175 |
outputs=[gr.Image(type="numpy", label="Annotated Output"), gr.Textbox(label="Emotion Analysis")],
|
| 176 |
title="Emotion Analysis",
|
| 177 |
+
description="Detects facial emotions using FER.",
|
| 178 |
+
live=True
|
| 179 |
)
|
| 180 |
|
| 181 |
objects_interface = gr.Interface(
|
|
|
|
| 183 |
inputs=gr.Image(sources=["webcam"], streaming=True, label="Capture the Scene"),
|
| 184 |
outputs=[gr.Image(type="numpy", label="Annotated Output"), gr.Textbox(label="Object Detection")],
|
| 185 |
title="Object Detection",
|
| 186 |
+
description="Detects objects using a pretrained Faster R-CNN.",
|
| 187 |
+
live=True
|
| 188 |
)
|
| 189 |
|
| 190 |
faces_interface = gr.Interface(
|
|
|
|
| 192 |
inputs=gr.Image(sources=["webcam"], streaming=True, label="Capture Your Face"),
|
| 193 |
outputs=[gr.Image(type="numpy", label="Annotated Output"), gr.Textbox(label="Face Detection")],
|
| 194 |
title="Face Detection",
|
| 195 |
+
description="Detects faces using MediaPipe.",
|
| 196 |
+
live=True
|
| 197 |
)
|
| 198 |
|
| 199 |
# -----------------------------
|