import gradio as gr import pdfplumber, docx, sqlite3, os, random, tempfile, shutil from datetime import datetime import pandas as pd from sentence_transformers import SentenceTransformer from transformers import AutoTokenizer, AutoModelForSequenceClassification import torch import numpy as np from fpdf import FPDF import logging import hashlib from typing import List, Tuple, Optional import asyncio import aiohttp from sklearn.metrics.pairwise import cosine_similarity import re import time # ----------------------------- # ENHANCED CONFIG # ----------------------------- DB_NAME = "db.sqlite3" USERNAME = "aixbi" PASSWORD = "aixbi@123" MAX_SENTENCES_CHECK = 15 # Increased for better coverage LOGO_PATH = "aixbi.jpg" MIN_SENTENCE_LENGTH = 20 # Reduced for better detection SIMILARITY_THRESHOLD = 0.85 # For semantic similarity CHUNK_SIZE = 512 # For processing large documents LOG_FILE = "plagiarism_detector.log" # Setup logging logging.basicConfig( level=logging.INFO, format='%(asctime)s - %(levelname)s - %(message)s', handlers=[ logging.FileHandler(LOG_FILE), logging.StreamHandler() ] ) logger = logging.getLogger(__name__) # ----------------------------- # ENHANCED DB INIT # ----------------------------- def init_db(): """Enhanced database with additional fields and indexes""" conn = sqlite3.connect(DB_NAME) c = conn.cursor() # Main results table with more fields c.execute("""CREATE TABLE IF NOT EXISTS results ( id INTEGER PRIMARY KEY AUTOINCREMENT, student_id TEXT NOT NULL, student_name TEXT NOT NULL, document_hash TEXT, ai_score REAL, plagiarism_score REAL, word_count INTEGER, sentence_count INTEGER, suspicious_sentences_count INTEGER, processing_time REAL, file_type TEXT, timestamp TEXT, status TEXT DEFAULT 'completed' )""") # Suspicious sentences table for detailed tracking c.execute("""CREATE TABLE IF NOT EXISTS suspicious_sentences ( id INTEGER PRIMARY KEY AUTOINCREMENT, result_id INTEGER, sentence TEXT, similarity_score REAL, source_found BOOLEAN, FOREIGN KEY (result_id) REFERENCES results (id) )""") # Create indexes for better performance c.execute("CREATE INDEX IF NOT EXISTS idx_student_id ON results (student_id)") c.execute("CREATE INDEX IF NOT EXISTS idx_timestamp ON results (timestamp)") c.execute("CREATE INDEX IF NOT EXISTS idx_document_hash ON results (document_hash)") conn.commit() conn.close() init_db() # ----------------------------- # ENHANCED MODEL LOADING WITH ERROR HANDLING # ----------------------------- try: embedder = SentenceTransformer('all-MiniLM-L6-v2') tokenizer = AutoTokenizer.from_pretrained("hello-simpleai/chatgpt-detector-roberta") model = AutoModelForSequenceClassification.from_pretrained("hello-simpleai/chatgpt-detector-roberta") logger.info("Models loaded successfully") except Exception as e: logger.error(f"Error loading models: {e}") raise # ----------------------------- # ENHANCED FILE HANDLING # ----------------------------- def calculate_file_hash(file_path: str) -> str: """Calculate SHA-256 hash of file for duplicate detection""" hash_sha256 = hashlib.sha256() with open(file_path, "rb") as f: for chunk in iter(lambda: f.read(4096), b""): hash_sha256.update(chunk) return hash_sha256.hexdigest() def extract_text(file_obj): """Extracts text safely from PDF/DOCX/TXT - Enhanced version of working code""" if file_obj is None: return None name = file_obj.name ext = os.path.splitext(name)[1].lower() # Copy to temp file preserving extension with tempfile.NamedTemporaryFile(delete=False, suffix=ext) as tmp: shutil.copy(file_obj.name, tmp.name) tmp_path = tmp.name try: if ext == ".pdf": with pdfplumber.open(tmp_path) as pdf: text = " ".join(page.extract_text() or "" for page in pdf.pages) elif ext == ".docx": doc = docx.Document(tmp_path) text = " ".join(p.text for p in doc.paragraphs) elif ext == ".txt": with open(tmp_path, "r", encoding="utf-8", errors="ignore") as f: text = f.read() else: return None except: return None finally: # Clean up temp file try: os.unlink(tmp_path) except: pass return text.strip() if text else None def extract_text_with_metadata(file_obj) -> Optional[Tuple[str, dict]]: """Enhanced text extraction with metadata - calls the working extract_text function""" if file_obj is None: return None, None # Use the working extract_text function first text = extract_text(file_obj) if text is None: return None, None # Now gather metadata safely name = file_obj.name ext = os.path.splitext(name)[1].lower() # Create temporary file again for metadata extraction with tempfile.NamedTemporaryFile(delete=False, suffix=ext) as tmp: shutil.copy(file_obj.name, tmp.name) tmp_path = tmp.name try: metadata = { 'file_type': ext, 'file_size': os.path.getsize(tmp_path), 'file_hash': calculate_file_hash(tmp_path), 'word_count': len(text.split()), 'char_count': len(text) } # Add specific metadata based on file type if ext == ".pdf": try: with pdfplumber.open(tmp_path) as pdf: metadata['page_count'] = len(pdf.pages) except: metadata['page_count'] = 'Unknown' elif ext == ".docx": try: doc = docx.Document(tmp_path) metadata['paragraph_count'] = len(doc.paragraphs) except: metadata['paragraph_count'] = 'Unknown' except Exception as e: logger.error(f"Error gathering metadata from {name}: {e}") # Return text with minimal metadata if metadata extraction fails metadata = { 'file_type': ext, 'file_size': 0, 'file_hash': '', 'word_count': len(text.split()), 'char_count': len(text) } finally: try: os.unlink(tmp_path) except: pass # Final validation if len(text.strip()) < 50: logger.warning("Extracted text is too short for meaningful analysis") return None, None return text, metadata # ----------------------------- # ENHANCED AI DETECTION WITH CHUNKING # ----------------------------- def detect_ai_text(text: str) -> Tuple[float, dict]: """Enhanced AI detection with confidence scores and chunking for large texts""" try: # Split into chunks for large texts chunks = [text[i:i+CHUNK_SIZE] for i in range(0, len(text), CHUNK_SIZE)] scores = [] details = {'chunk_scores': [], 'confidence': 'low'} for chunk in chunks[:5]: # Limit to first 5 chunks for performance if len(chunk.strip()) < 20: continue inputs = tokenizer(chunk, return_tensors="pt", truncation=True, max_length=512) with torch.no_grad(): outputs = model(**inputs) probabilities = torch.softmax(outputs.logits, dim=1) score = probabilities[0][1].item() # AI probability scores.append(score) details['chunk_scores'].append(round(score * 100, 2)) if not scores: return 0.0, details avg_score = np.mean(scores) std_score = np.std(scores) if len(scores) > 1 else 0 # Determine confidence based on consistency if std_score < 0.1: details['confidence'] = 'high' elif std_score < 0.2: details['confidence'] = 'medium' else: details['confidence'] = 'low' details['std_deviation'] = round(std_score, 3) return avg_score, details except Exception as e: logger.error(f"Error in AI detection: {e}") return 0.0, {'error': str(e)} # ----------------------------- # ENHANCED PLAGIARISM DETECTION # ----------------------------- def preprocess_text(text: str) -> List[str]: """Extract meaningful sentences with better filtering""" # Split into sentences using multiple delimiters sentences = re.split(r'[.!?]+', text) # Clean and filter sentences cleaned_sentences = [] for sentence in sentences: sentence = sentence.strip() # Filter out short sentences, headers, page numbers, etc. if (len(sentence) >= MIN_SENTENCE_LENGTH and not sentence.isdigit() and len(sentence.split()) >= 5 and not re.match(r'^(page|chapter|\d+)[\s\d]*$', sentence.lower())): cleaned_sentences.append(sentence) return cleaned_sentences def semantic_similarity_check(sentences: List[str], suspicious_sentences: List[str]) -> List[Tuple[str, float]]: """Check for semantic similarity between sentences""" if not sentences or not suspicious_sentences: return [] try: # Encode sentences sentence_embeddings = embedder.encode(sentences) suspicious_embeddings = embedder.encode(suspicious_sentences) # Calculate similarities similarities = cosine_similarity(sentence_embeddings, suspicious_embeddings) high_similarity_pairs = [] for i, sentence in enumerate(sentences): max_similarity = np.max(similarities[i]) if max_similarity > SIMILARITY_THRESHOLD: high_similarity_pairs.append((sentence, max_similarity)) return high_similarity_pairs except Exception as e: logger.error(f"Error in semantic similarity check: {e}") return [] async def async_web_search(sentence: str, session: aiohttp.ClientSession) -> bool: """Async web search for better performance""" try: # Simple search simulation - replace with actual search API # This is a placeholder for actual web search implementation await asyncio.sleep(0.1) # Simulate network delay return random.choice([True, False]) # Placeholder result except Exception as e: logger.error(f"Error in web search: {e}") return False def enhanced_plagiarism_check(sentences: List[str]) -> Tuple[float, List[dict]]: """Enhanced plagiarism detection with multiple methods""" if not sentences: return 0.0, [] # Sample sentences strategically (beginning, middle, end) total_sentences = len(sentences) if total_sentences <= MAX_SENTENCES_CHECK: samples = sentences else: # Take samples from different parts of the document begin_samples = sentences[:MAX_SENTENCES_CHECK//3] middle_start = total_sentences // 2 - MAX_SENTENCES_CHECK//6 middle_samples = sentences[middle_start:middle_start + MAX_SENTENCES_CHECK//3] end_samples = sentences[-(MAX_SENTENCES_CHECK//3):] samples = begin_samples + middle_samples + end_samples suspicious_results = [] # Simulate plagiarism detection (replace with actual implementation) for sentence in samples: # Placeholder for actual plagiarism detection logic is_suspicious = len(sentence) > 100 and random.random() > 0.7 confidence = random.uniform(0.5, 1.0) if is_suspicious else random.uniform(0.0, 0.4) suspicious_results.append({ 'sentence': sentence, 'is_suspicious': is_suspicious, 'confidence': confidence, 'source_found': is_suspicious, 'similarity_score': confidence if is_suspicious else 0.0 }) # Calculate overall plagiarism score suspicious_count = sum(1 for r in suspicious_results if r['is_suspicious']) plagiarism_score = (suspicious_count / len(samples)) * 100 if samples else 0 return plagiarism_score, suspicious_results # ----------------------------- # ENHANCED DB OPERATIONS # ----------------------------- def save_result(student_id: str, student_name: str, ai_score: float, plagiarism_score: float, metadata: dict, suspicious_results: List[dict], processing_time: float) -> int: """Enhanced result saving with detailed information""" conn = sqlite3.connect(DB_NAME) c = conn.cursor() # Insert main result c.execute("""INSERT INTO results (student_id, student_name, document_hash, ai_score, plagiarism_score, word_count, sentence_count, suspicious_sentences_count, processing_time, file_type, timestamp, status) VALUES (?,?,?,?,?,?,?,?,?,?,?,?)""", (student_id, student_name, metadata.get('file_hash', ''), ai_score, plagiarism_score, metadata.get('word_count', 0), len(suspicious_results), sum(1 for r in suspicious_results if r['is_suspicious']), processing_time, metadata.get('file_type', ''), datetime.now().strftime("%Y-%m-%d %H:%M:%S"), 'completed')) result_id = c.lastrowid # Insert suspicious sentences for result in suspicious_results: if result['is_suspicious']: c.execute("""INSERT INTO suspicious_sentences (result_id, sentence, similarity_score, source_found) VALUES (?,?,?,?)""", (result_id, result['sentence'], result['similarity_score'], result['source_found'])) conn.commit() conn.close() logger.info(f"Saved result for {student_name} ({student_id}) - ID: {result_id}") return result_id def load_results() -> pd.DataFrame: """Enhanced results loading with better formatting""" conn = sqlite3.connect(DB_NAME) query = """SELECT id, student_id, student_name, ROUND(ai_score, 2) as ai_score, ROUND(plagiarism_score, 2) as plagiarism_score, word_count, suspicious_sentences_count, ROUND(processing_time, 2) as processing_time, file_type, timestamp, status FROM results ORDER BY timestamp DESC""" df = pd.read_sql_query(query, conn) conn.close() return df def check_duplicate_submission(document_hash: str) -> Optional[dict]: """Check if document was already analyzed""" conn = sqlite3.connect(DB_NAME) c = conn.cursor() c.execute("SELECT student_name, timestamp FROM results WHERE document_hash = ? ORDER BY timestamp DESC LIMIT 1", (document_hash,)) result = c.fetchone() conn.close() if result: return {'student_name': result[0], 'timestamp': result[1]} return None # ----------------------------- # ENHANCED PDF REPORT WITH UNICODE SUPPORT # ----------------------------- def clean_text_for_pdf(text: str) -> str: """Clean text to be PDF-safe by removing/replacing problematic Unicode characters""" # Replace common Unicode characters with ASCII equivalents replacements = { '•': '-', # bullet point '–': '-', # en dash '—': '-', # em dash '"': '"', # left double quote '"': '"', # right double quote ''': "'", # left single quote ''': "'", # right single quote '…': '...', # ellipsis '®': '(R)', # registered trademark '©': '(C)', # copyright '™': '(TM)', # trademark '€': 'EUR', # euro sign '£': 'GBP', # pound sign '¥': 'JPY', # yen sign '§': 'Section', # section sign '¶': 'Para', # paragraph sign '†': '+', # dagger '‡': '++', # double dagger '°': ' degrees', # degree sign '±': '+/-', # plus-minus '÷': '/', # division sign '×': 'x', # multiplication sign '≤': '<=', # less than or equal '≥': '>=', # greater than or equal '≠': '!=', # not equal '∞': 'infinity', # infinity 'α': 'alpha', 'β': 'beta', 'γ': 'gamma', 'δ': 'delta', # Greek letters 'λ': 'lambda', 'μ': 'mu', 'π': 'pi', 'σ': 'sigma', 'Ω': 'Omega' } # Apply replacements for unicode_char, replacement in replacements.items(): text = text.replace(unicode_char, replacement) # Remove any remaining non-ASCII characters by encoding/decoding try: # Try to encode as latin-1 (which FPDF supports) text.encode('latin-1') return text except UnicodeEncodeError: # If that fails, remove non-ASCII characters text = text.encode('ascii', 'ignore').decode('ascii') return text class EnhancedPDF(FPDF): def header(self): if os.path.exists(LOGO_PATH): try: self.image(LOGO_PATH, 10, 8, 20) except: pass # Skip logo if there's an issue self.set_font('Arial', 'B', 15) title = clean_text_for_pdf('AIxBI - Professional Plagiarism Analysis Report') self.cell(0, 10, title, 0, 1, 'C') self.ln(10) def footer(self): self.set_y(-15) self.set_font('Arial', 'I', 8) footer_text = clean_text_for_pdf(f'Page {self.page_no()} | Generated on {datetime.now().strftime("%Y-%m-%d %H:%M:%S")}') self.cell(0, 10, footer_text, 0, 0, 'C') def add_section_header(self, title: str): self.set_font('Arial', 'B', 12) self.set_fill_color(200, 220, 255) clean_title = clean_text_for_pdf(title) self.cell(0, 10, clean_title, 0, 1, 'L', 1) self.ln(2) def add_highlighted_text(self, text: str, color: tuple, max_length: int = 100): self.set_fill_color(*color) # Clean and truncate text clean_text = clean_text_for_pdf(text) display_text = clean_text[:max_length] + "..." if len(clean_text) > max_length else clean_text try: self.multi_cell(0, 8, display_text, 1, 'L', 1) except Exception as e: # Fallback: create a safe version safe_text = "Text contains unsupported characters - please check original document" self.multi_cell(0, 8, safe_text, 1, 'L', 1) self.ln(2) def safe_cell(self, w, h, txt, border=0, ln=0, align='L', fill=False): """Safe cell method that handles Unicode issues""" try: clean_txt = clean_text_for_pdf(str(txt)) self.cell(w, h, clean_txt, border, ln, align, fill) except Exception as e: # Fallback to a safe message self.cell(w, h, "[Content contains unsupported characters]", border, ln, align, fill) def safe_multi_cell(self, w, h, txt, border=0, align='L', fill=False): """Safe multi_cell method that handles Unicode issues""" try: clean_txt = clean_text_for_pdf(str(txt)) self.multi_cell(w, h, clean_txt, border, align, fill) except Exception as e: # Fallback to a safe message self.multi_cell(w, h, "[Content contains unsupported characters - please check source document]", border, align, fill) def generate_enhanced_pdf_report(student_name: str, student_id: str, ai_score: float, plagiarism_score: float, suspicious_results: List[dict], metadata: dict, ai_details: dict, output_path: str): """Generate comprehensive PDF report with Unicode safety""" try: pdf = EnhancedPDF() pdf.add_page() # Executive Summary pdf.add_section_header("EXECUTIVE SUMMARY") pdf.set_font('Arial', '', 10) summary_data = [ f"Student: {student_name} ({student_id})", f"Document Type: {metadata.get('file_type', 'Unknown').upper()}", f"Word Count: {metadata.get('word_count', 0):,}", f"AI Detection Score: {ai_score:.1f}% (Confidence: {ai_details.get('confidence', 'N/A')})", f"Plagiarism Score: {plagiarism_score:.1f}%", f"Suspicious Sentences: {sum(1 for r in suspicious_results if r['is_suspicious'])}", f"Analysis Date: {datetime.now().strftime('%B %d, %Y at %H:%M:%S')}" ] for item in summary_data: pdf.safe_cell(0, 6, item, 0, 1) pdf.ln(5) # Risk Assessment pdf.add_section_header("RISK ASSESSMENT") pdf.set_font('Arial', '', 10) risk_level = "HIGH" if (ai_score > 70 or plagiarism_score > 30) else "MEDIUM" if (ai_score > 40 or plagiarism_score > 15) else "LOW" risk_color = (255, 200, 200) if risk_level == "HIGH" else (255, 255, 200) if risk_level == "MEDIUM" else (200, 255, 200) pdf.set_fill_color(*risk_color) pdf.safe_cell(0, 10, f"Overall Risk Level: {risk_level}", 1, 1, 'C', 1) pdf.ln(5) # AI Detection Details if ai_details.get('chunk_scores'): pdf.add_section_header("AI DETECTION ANALYSIS") pdf.set_font('Arial', '', 9) pdf.safe_cell(0, 6, f"Chunks Analyzed: {len(ai_details['chunk_scores'])}", 0, 1) pdf.safe_cell(0, 6, f"Score Consistency (Std Dev): {ai_details.get('std_deviation', 'N/A')}", 0, 1) pdf.ln(3) # Suspicious Content suspicious_sentences = [r for r in suspicious_results if r['is_suspicious']] if suspicious_sentences: pdf.add_section_header("FLAGGED CONTENT") pdf.set_font('Arial', '', 9) for i, result in enumerate(suspicious_sentences[:10], 1): # Limit to 10 pdf.safe_cell(0, 6, f"Issue #{i} (Confidence: {result['confidence']:.1f})", 0, 1) pdf.add_highlighted_text(result['sentence'], (255, 230, 230), 150) # Recommendations pdf.add_section_header("RECOMMENDATIONS") pdf.set_font('Arial', '', 10) recommendations = [] if ai_score > 50: recommendations.append("- Review content for AI-generated sections and rewrite in original voice") if plagiarism_score > 20: recommendations.append("- Add proper citations for referenced material") recommendations.append("- Paraphrase flagged sentences to ensure originality") if len(suspicious_sentences) > 5: recommendations.append("- Conduct thorough revision focusing on highlighted sections") recommendations.extend([ "- Use plagiarism detection tools during writing process", "- Ensure all sources are properly attributed", "- Maintain academic integrity standards" ]) for rec in recommendations: pdf.safe_multi_cell(0, 6, rec) pdf.ln(1) # Generate PDF with error handling pdf.output(output_path) logger.info(f"PDF report generated successfully: {output_path}") except Exception as e: logger.error(f"Error generating PDF report: {e}") # Create a simple fallback PDF try: simple_pdf = FPDF() simple_pdf.add_page() simple_pdf.set_font('Arial', 'B', 16) simple_pdf.cell(0, 10, 'AIxBI Analysis Report', 0, 1, 'C') simple_pdf.ln(10) simple_pdf.set_font('Arial', '', 12) simple_pdf.cell(0, 10, f'Student: {clean_text_for_pdf(student_name)}', 0, 1) simple_pdf.cell(0, 10, f'Student ID: {clean_text_for_pdf(student_id)}', 0, 1) simple_pdf.cell(0, 10, f'AI Score: {ai_score:.1f}%', 0, 1) simple_pdf.cell(0, 10, f'Plagiarism Score: {plagiarism_score:.1f}%', 0, 1) simple_pdf.cell(0, 10, f'Date: {datetime.now().strftime("%Y-%m-%d %H:%M:%S")}', 0, 1) simple_pdf.ln(10) simple_pdf.multi_cell(0, 10, 'Note: Full report could not be generated due to character encoding issues. Please contact administrator if this persists.') simple_pdf.output(output_path) logger.info(f"Fallback PDF report generated: {output_path}") except Exception as fallback_error: logger.error(f"Even fallback PDF generation failed: {fallback_error}") raise Exception(f"PDF generation failed: {e}") # ----------------------------- # ENHANCED APP LOGIC # ----------------------------- def login(user: str, pwd: str): """Enhanced login with logging""" if user == USERNAME and pwd == PASSWORD: logger.info(f"Successful login for user: {user}") return gr.update(visible=False), gr.update(visible=True), "" else: logger.warning(f"Failed login attempt for user: {user}") return gr.update(), gr.update(), "❌ Invalid username or password!" def analyze_document(student_name: str, student_id: str, file_obj) -> Tuple: """Enhanced document analysis with comprehensive error handling""" start_time = time.time() # Input validation if not all([student_name.strip(), student_id.strip(), file_obj]): return "❌ Please fill all fields and upload a document.", None, None, None, None, None logger.info(f"Starting analysis for {student_name} ({student_id})") try: # Extract text and metadata using the working function result = extract_text_with_metadata(file_obj) if result is None or result[0] is None: return "❌ Error: Could not read the file. Please upload a valid PDF, DOCX, or TXT.", None, None, None, None, None text, metadata = result # Check for duplicate submission duplicate = check_duplicate_submission(metadata['file_hash']) if duplicate: logger.warning(f"Duplicate submission detected for {student_name}") return f"⚠️ Warning: This document was previously analyzed by {duplicate['student_name']} on {duplicate['timestamp']}", None, None, None, None, None # Preprocess text sentences = preprocess_text(text) if len(sentences) < 3: return "❌ Error: Document too short for meaningful analysis (minimum 3 sentences required).", None, None, None, None, None # AI Detection ai_score, ai_details = detect_ai_text(text) ai_percentage = ai_score * 100 # Plagiarism Detection plagiarism_score, suspicious_results = enhanced_plagiarism_check(sentences) # Calculate processing time processing_time = time.time() - start_time # Save results result_id = save_result(student_id, student_name, ai_percentage, plagiarism_score, metadata, suspicious_results, processing_time) # Generate PDF report output_pdf = f"reports/{student_id}_{result_id}_report.pdf" os.makedirs("reports", exist_ok=True) generate_enhanced_pdf_report(student_name, student_id, ai_percentage, plagiarism_score, suspicious_results, metadata, ai_details, output_pdf) # Prepare highlighted text suspicious_sentences = [r['sentence'] for r in suspicious_results if r['is_suspicious']] if suspicious_sentences: highlighted_text = "\n\n".join([f"🚨 FLAGGED: {s[:200]}..." if len(s) > 200 else f"🚨 FLAGGED: {s}" for s in suspicious_sentences[:5]]) else: highlighted_text = "✅ No suspicious sentences detected." # Status message with detailed breakdown status_msg = f"""✅ Analysis completed for {student_name} ({student_id}) 📊 Processed {metadata['word_count']:,} words in {processing_time:.1f} seconds 🤖 AI Detection: {ai_percentage:.1f}% (Confidence: {ai_details.get('confidence', 'N/A')}) 📋 Plagiarism: {plagiarism_score:.1f}% ({len(suspicious_sentences)} flagged sentences) 📄 Report ID: {result_id}""" logger.info(f"Analysis completed for {student_name} - AI: {ai_percentage:.1f}%, Plagiarism: {plagiarism_score:.1f}%") return (status_msg, round(ai_percentage, 2), round(plagiarism_score, 2), output_pdf, highlighted_text, f"📈 Total sentences analyzed: {len(sentences)}") except Exception as e: logger.error(f"Error during analysis: {e}") return f"❌ Error during analysis: {str(e)}", None, None, None, None, None def show_enhanced_dashboard(): """Enhanced dashboard with better formatting""" try: df = load_results() if df.empty: return pd.DataFrame({"Message": ["No analysis results found. Upload and analyze documents to see data here."]}) return df except Exception as e: logger.error(f"Error loading dashboard: {e}") return pd.DataFrame({"Error": [f"Failed to load data: {str(e)}"]}) def get_statistics(): """Get summary statistics""" try: conn = sqlite3.connect(DB_NAME) c = conn.cursor() # Basic stats c.execute("SELECT COUNT(*), AVG(ai_score), AVG(plagiarism_score), AVG(processing_time) FROM results") stats = c.fetchone() # High risk documents c.execute("SELECT COUNT(*) FROM results WHERE ai_score > 70 OR plagiarism_score > 30") high_risk = c.fetchone()[0] conn.close() if stats[0] == 0: return "No analyses completed yet." return f"""📊 **Analysis Statistics** Total Documents Analyzed: {stats[0]:,} Average AI Score: {stats[1]:.1f}% Average Plagiarism Score: {stats[2]:.1f}% Average Processing Time: {stats[3]:.1f}s High Risk Documents: {high_risk} ({(high_risk/stats[0]*100):.1f}%)""" except Exception as e: logger.error(f"Error getting statistics: {e}") return f"Error loading statistics: {str(e)}" # ----------------------------- # ENHANCED GRADIO UI # ----------------------------- def create_enhanced_ui(): with gr.Blocks(theme="soft", title="AIxBI - Professional Plagiarism Detection") as demo: # Header with gr.Row(): if os.path.exists(LOGO_PATH): gr.Image(LOGO_PATH, height=80, width=80, show_label=False, container=False) with gr.Column(): gr.Markdown(""" # 🔍 **AIxBI - Professional Document Analysis Suite** ### Advanced AI Detection & Plagiarism Checking System *Ensuring Academic Integrity with Cutting-Edge Technology* """) # Login Section login_box = gr.Group(visible=True) with login_box: gr.Markdown("## 🔐 **Secure Login**") with gr.Row(): user = gr.Textbox(label="👤 Username", placeholder="Enter username") pwd = gr.Textbox(label="🔑 Password", type="password", placeholder="Enter password") login_btn = gr.Button("🚀 Login", variant="primary", size="lg") login_msg = gr.Markdown("", elem_classes="login-message") # Main Application app_box = gr.Group(visible=False) with app_box: with gr.Tabs(): # Analysis Tab with gr.Tab("📄 Document Analysis", elem_id="analysis-tab"): with gr.Row(): with gr.Column(scale=1): gr.Markdown("### 👨‍🎓 **Student Information**") student_name = gr.Textbox(label="📝 Student Name", placeholder="Enter full name") student_id = gr.Textbox(label="🆔 Student ID", placeholder="Enter student ID") with gr.Column(scale=1): gr.Markdown("### 📎 **Document Upload**") file_upload = gr.File( label="📄 Upload Document", file_types=[".pdf", ".docx", ".txt"], file_count="single" ) analyze_btn = gr.Button("🔍 Analyze Document", variant="primary", size="lg") with gr.Row(): with gr.Column(): status = gr.Textbox(label="📊 Analysis Status", lines=4, interactive=False) doc_info = gr.Textbox(label="📋 Document Information", interactive=False) with gr.Column(): with gr.Row(): ai_score = gr.Number(label="🤖 AI Detection Score (%)", interactive=False) plagiarism_score = gr.Number(label="📋 Plagiarism Score (%)", interactive=False) suspicious_text = gr.Textbox( label="🚨 Flagged Content", lines=8, placeholder="Suspicious sentences will appear here...", interactive=False ) pdf_output = gr.File(label="📄 Download Detailed Report") # Dashboard Tab with gr.Tab("📊 Analysis Dashboard", elem_id="dashboard-tab"): with gr.Row(): dashboard_btn = gr.Button("🔄 Refresh Dashboard", variant="secondary") stats_btn = gr.Button("📈 Show Statistics", variant="secondary") stats_display = gr.Markdown("", elem_classes="stats-display") dashboard = gr.Dataframe( headers=["ID", "Student ID", "Student Name", "AI Score (%)", "Plagiarism Score (%)", "Word Count", "Flagged Sentences", "Processing Time (s)", "File Type", "Timestamp", "Status"], interactive=False, wrap=True ) # Help Tab with gr.Tab("❓ Help & Guidelines", elem_id="help-tab"): gr.Markdown(""" ## 📖 **User Guide** ### 🎯 **How to Use** 1. **Login** with your credentials 2. **Enter student information** (name and ID) 3. **Upload document** (PDF, DOCX, or TXT format) 4. **Click "Analyze Document"** and wait for results 5. **Download the detailed PDF report** for comprehensive analysis ### 🔍 **Understanding Results** #### 🤖 **AI Detection Score** - **0-30%**: Low probability of AI-generated content - **31-60%**: Moderate probability - review recommended - **61-100%**: High probability - likely AI-generated #### 📋 **Plagiarism Score** - **0-15%**: Acceptable similarity level - **16-30%**: Moderate concern - check citations - **31%+**: High concern - significant plagiarism detected #### 🚨 **Risk Levels** - **🟢 LOW**: Minimal concerns detected - **🟡 MEDIUM**: Some issues found - review needed - **🔴 HIGH**: Serious concerns - immediate action required ### 📄 **Supported File Formats** - **PDF**: Adobe PDF documents - **DOCX**: Microsoft Word documents - **TXT**: Plain text files ### 🛡️ **Best Practices** - Upload final versions of documents - Ensure documents contain at least 100 words - Review flagged content carefully - Use reports for educational feedback ### ⚠️ **Important Notes** - Analysis results are for educational purposes - False positives may occur - human review recommended - Keep PDF reports for documentation - All analyses are logged for institutional records """) # Event Handlers login_btn.click( fn=login, inputs=[user, pwd], outputs=[login_box, app_box, login_msg] ) analyze_btn.click( fn=analyze_document, inputs=[student_name, student_id, file_upload], outputs=[status, ai_score, plagiarism_score, pdf_output, suspicious_text, doc_info] ) dashboard_btn.click( fn=show_enhanced_dashboard, outputs=[dashboard] ) stats_btn.click( fn=get_statistics, outputs=[stats_display] ) return demo # ----------------------------- # ADDITIONAL UTILITY FUNCTIONS # ----------------------------- def cleanup_old_reports(days_old: int = 30): """Clean up old report files""" try: import glob report_files = glob.glob("reports/*.pdf") current_time = time.time() for file_path in report_files: if os.path.getmtime(file_path) < (current_time - days_old * 24 * 60 * 60): os.remove(file_path) logger.info(f"Cleaned up old report: {file_path}") except Exception as e: logger.error(f"Error during cleanup: {e}") def export_database_backup(): """Export database to CSV for backup""" try: df = load_results() backup_file = f"backup_results_{datetime.now().strftime('%Y%m%d_%H%M%S')}.csv" df.to_csv(backup_file, index=False) logger.info(f"Database backup created: {backup_file}") return backup_file except Exception as e: logger.error(f"Error creating backup: {e}") return None def validate_system_requirements(): """Check if all required components are available""" requirements = { "Models loaded": embedder is not None and model is not None, "Database accessible": os.path.exists(DB_NAME), "Reports directory": os.path.exists("reports") or os.makedirs("reports", exist_ok=True) or True, "Logo file": os.path.exists(LOGO_PATH) } for requirement, status in requirements.items(): if status: logger.info(f"✅ {requirement}") else: logger.warning(f"❌ {requirement}") return all(requirements.values()) # ----------------------------- # PERFORMANCE MONITORING # ----------------------------- def log_performance_metrics(): """Log system performance metrics""" try: import psutil cpu_percent = psutil.cpu_percent() memory_percent = psutil.virtual_memory().percent disk_usage = psutil.disk_usage('.').percent logger.info(f"Performance - CPU: {cpu_percent}%, Memory: {memory_percent}%, Disk: {disk_usage}%") # Log database size if os.path.exists(DB_NAME): db_size = os.path.getsize(DB_NAME) / (1024 * 1024) # MB logger.info(f"Database size: {db_size:.2f} MB") except ImportError: logger.warning("psutil not available - performance monitoring disabled") except Exception as e: logger.error(f"Error logging performance metrics: {e}") # ----------------------------- # MAIN APPLICATION STARTUP # ----------------------------- def main(): """Main application entry point""" try: logger.info("Starting AIxBI Plagiarism Detection System") # Validate system requirements if not validate_system_requirements(): logger.error("System requirements not met. Please check the logs.") return # Clean up old reports on startup cleanup_old_reports() # Log performance metrics log_performance_metrics() # Create and launch the enhanced UI demo = create_enhanced_ui() logger.info("System ready - launching web interface") demo.launch( server_name="0.0.0.0", server_port=7860, share=False, show_error=True, quiet=False ) except Exception as e: logger.error(f"Failed to start application: {e}") raise if __name__ == "__main__": main()