tstone87's picture
Update app.py
7aa805e verified
raw
history blame
6.75 kB
import sys
import gradio as gr
import os
import tempfile
import cv2
import requests
from ultralytics import YOLO
# Remove extra CLI arguments that Spaces might pass
sys.argv = [arg for arg in sys.argv if arg != "--import"]
# Load the YOLO11-pose model
model = YOLO("yolo11n-pose.pt")
def process_input(uploaded_file, youtube_link, image_url, sensitivity):
"""
Process input from Upload, YouTube, or Image URL.
Priority: YouTube link > Image URL > Uploaded file.
Sensitivity is the confidence threshold.
"""
input_path = None
temp_files = []
# Priority 1: YouTube link
if youtube_link and youtube_link.strip():
try:
from pytubefix import YouTube # Use pytubefix instead of pytube
yt = YouTube(youtube_link)
stream = yt.streams.filter(file_extension='mp4', progressive=True).order_by("resolution").desc().first()
if not stream:
return None, None, None, "No suitable mp4 stream found."
temp_path = os.path.join(tempfile.gettempdir(), f"yt_{os.urandom(8).hex()}.mp4")
stream.download(output_path=tempfile.gettempdir(), filename=os.path.basename(temp_path))
input_path = temp_path
temp_files.append(input_path)
except Exception as e:
return None, None, None, f"Error downloading YouTube video: {str(e)}"
# Priority 2: Image URL
elif image_url and image_url.strip():
try:
response = requests.get(image_url, stream=True, timeout=10)
response.raise_for_status()
temp_path = os.path.join(tempfile.gettempdir(), f"img_{os.urandom(8).hex()}.jpg")
with open(temp_path, "wb") as f:
f.write(response.content)
input_path = temp_path
temp_files.append(input_path)
except Exception as e:
return None, None, None, f"Error downloading image: {str(e)}"
# Priority 3: Uploaded file
elif uploaded_file is not None:
input_path = uploaded_file.name
else:
return None, None, None, "Please provide an input."
# Process the file
ext = os.path.splitext(input_path)[1].lower()
video_exts = [".mp4", ".mov", ".avi", ".webm"]
output_path = None
try:
if ext in video_exts:
# Video processing
cap = cv2.VideoCapture(input_path)
if not cap.isOpened():
return None, None, None, "Error opening video file."
fps = cap.get(cv2.CAP_PROP_FPS)
width = int(cap.get(cv2.CAP_PROP_FRAME_WIDTH))
height = int(cap.get(cv2.CAP_PROP_FRAME_HEIGHT))
# Create output video
output_path = os.path.join(tempfile.gettempdir(), f"out_{os.urandom(8).hex()}.mp4")
fourcc = cv2.VideoWriter_fourcc(*'mp4v')
out = cv2.VideoWriter(output_path, fourcc, fps, (width, height))
while True:
ret, frame = cap.read()
if not ret:
break
# Convert BGR to RGB for YOLO
frame_rgb = cv2.cvtColor(frame, cv2.COLOR_BGR2RGB)
results = model.predict(source=frame_rgb, conf=sensitivity)[0]
annotated_frame = results.plot()
# Convert back to BGR for video writing
annotated_frame_bgr = cv2.cvtColor(annotated_frame, cv2.COLOR_RGB2BGR)
out.write(annotated_frame_bgr)
cap.release()
out.release()
temp_files.append(output_path)
if os.path.getsize(output_path) == 0:
return None, None, None, "Error: Output video is empty."
return output_path, None, output_path, "Video processed successfully!"
else:
# Image processing
results = model.predict(source=input_path, conf=sensitivity)[0]
annotated = results.plot()
output_path = os.path.join(tempfile.gettempdir(), f"out_{os.urandom(8).hex()}.jpg")
cv2.imwrite(output_path, annotated)
temp_files.append(output_path)
return output_path, output_path, None, "Image processed successfully!"
except Exception as e:
return None, None, None, f"Processing error: {str(e)}"
finally:
# Clean up temporary input files (but keep output for download)
for f in temp_files[:-1]: # Exclude output_path
if f and os.path.exists(f):
try:
os.remove(f)
except:
pass
# Gradio interface remains mostly the same
with gr.Blocks(css="""
.result_img > img {
width: 100%;
height: auto;
object-fit: contain;
}
""") as demo:
with gr.Row():
with gr.Column(scale=1):
gr.HTML("<div style='text-align:center;'><img src='https://huggingface.co/spaces/tstone87/stance-detection/resolve/main/crowdresult.jpg' style='width:25%;'/></div>")
gr.Markdown("## Pose Detection with YOLO11-pose")
with gr.Tabs():
with gr.TabItem("Upload File"):
file_input = gr.File(label="Upload Image/Video")
with gr.TabItem("YouTube Link"):
youtube_input = gr.Textbox(label="YouTube Link", placeholder="https://...")
with gr.TabItem("Image URL"):
image_url_input = gr.Textbox(label="Image URL", placeholder="https://...")
sensitivity_slider = gr.Slider(minimum=0.0, maximum=1.0, step=0.01, value=0.2,
label="Sensitivity (Confidence Threshold)")
with gr.Column(scale=2):
output_image = gr.Image(label="Annotated Output (Image)", elem_classes="result_img")
output_video = gr.Video(label="Annotated Output (Video)")
output_file = gr.File(label="Download Annotated Output")
output_text = gr.Textbox(label="Status", interactive=False)
file_input.change(
fn=process_input,
inputs=[file_input, gr.State(""), gr.State(""), sensitivity_slider],
outputs=[output_file, output_image, output_video, output_text]
)
youtube_input.change(
fn=process_input,
inputs=[gr.State(None), youtube_input, gr.State(""), sensitivity_slider],
outputs=[output_file, output_image, output_video, output_text]
)
image_url_input.change(
fn=process_input,
inputs=[gr.State(None), gr.State(""), image_url_input, sensitivity_slider],
outputs=[output_file, output_image, output_video, output_text]
)
if __name__ == "__main__":
demo.launch()