ShallowCodeResearch / README.md
HallD's picture
Upload 60 files
df2b222 verified
|
raw
history blame
6.07 kB
---
title: ShallowCodeResearch
emoji: πŸ“‰
colorFrom: blue
colorTo: pink
sdk: gradio
sdk_version: 5.33.0
app_file: app.py
pinned: false
short_description: Coding research assistant that generates code and tests it
tags:
- mcp
- multi-agent
- research
- code-generation
- ai-assistant
- gradio
- python
- web-search
- llm
- modal
python_version: "3.12"
---
---
# MCP Hub - Multi-Agent AI Research & Code Assistant
πŸš€ **Advanced multi-agent system for AI-powered research and code generation**
## What is MCP Hub?
MCP Hub is a sophisticated multi-agent research and code assistant built using Gradio's Model Context Protocol (MCP) server functionality. It orchestrates specialized AI agents to provide comprehensive research capabilities and generate executable Python code.
## ✨ Key Features
- 🧠 **Multi-Agent Architecture**: Specialized agents working in orchestrated workflows
- πŸ” **Intelligent Research**: Web search with automatic summarization and citation formatting
- πŸ’» **Code Generation**: Context-aware Python code creation with secure execution
- πŸ”— **MCP Server**: Built-in MCP server for seamless agent communication
- 🎯 **Multiple LLM Support**: Compatible with Nebius, OpenAI, Anthropic, and HuggingFace
- πŸ›‘οΈ **Secure Execution**: Modal sandbox environment for safe code execution
- πŸ“Š **Performance Monitoring**: Advanced metrics collection and health monitoring
## πŸš€ Quick Start
1. **Configure your environment** by setting up API keys in the Settings tab
2. **Choose your LLM provider** (Nebius recommended for best performance)
3. **Input your research query** in the Orchestrator Flow tab
4. **Watch the magic happen** as agents collaborate to research and generate code
## πŸ—οΈ Architecture
### Core Agents
- **Question Enhancer**: Breaks down complex queries into focused sub-questions
- **Web Search Agent**: Performs targeted searches using Tavily API
- **LLM Processor**: Handles text processing, summarization, and analysis
- **Citation Formatter**: Manages academic citation formatting (APA style)
- **Code Generator**: Creates contextually-aware Python code
- **Code Runner**: Executes code in secure Modal sandboxes
- **Orchestrator**: Coordinates the complete workflow
### Workflow Example
```
User Query: "Create Python code to analyze Twitter sentiment"
↓
Question Enhancement: Split into focused sub-questions
↓
Web Research: Search for Twitter APIs, sentiment libraries, examples
↓
Context Integration: Combine research into comprehensive context
↓
Code Generation: Create executable Python script
↓
Secure Execution: Run code in Modal sandbox
↓
Results: Code + output + research summary + citations
```
## πŸ› οΈ Setup Requirements
### Required API Keys
- **LLM Provider** (choose one):
- Nebius API (recommended)
- OpenAI API
- Anthropic API
- HuggingFace Inference API
- **Tavily API** (for web search)
- **Modal Account** (for code execution)
### Environment Configuration
Set these environment variables or configure in the app:
```bash
LLM_PROVIDER=nebius # Your chosen provider
NEBIUS_API_KEY=your_key_here
TAVILY_API_KEY=your_key_here
# Modal setup handled automatically
```
## 🎯 Use Cases
### Research & Development
- **Academic Research**: Automated literature review and citation management
- **Technical Documentation**: Generate comprehensive guides with current information
- **Market Analysis**: Research trends and generate analytical reports
### Code Generation
- **Prototype Development**: Rapidly create functional code based on requirements
- **API Integration**: Generate code for working with various APIs and services
- **Data Analysis**: Create scripts for data processing and visualization
### Learning & Education
- **Code Examples**: Generate educational code samples with explanations
- **Concept Exploration**: Research and understand complex programming concepts
- **Best Practices**: Learn current industry standards and methodologies
## πŸ”§ Advanced Features
### Performance Monitoring
- Real-time metrics collection
- Response time tracking
- Success rate monitoring
- Resource usage analytics
### Intelligent Caching
- Reduces redundant API calls
- Improves response times
- Configurable TTL settings
### Fault Tolerance
- Circuit breaker protection
- Rate limiting management
- Graceful error handling
- Automatic retry mechanisms
### Sandbox Pool Management
- Pre-warmed execution environments
- Optimized performance
- Resource pooling
- Automatic scaling
## πŸ“± Interface Tabs
1. **Orchestrator Flow**: Complete end-to-end workflow
2. **Individual Agents**: Access each agent separately for specific tasks
3. **Advanced Features**: System monitoring and performance analytics
## 🀝 MCP Integration
This application demonstrates advanced MCP (Model Context Protocol) implementation:
- **Server Architecture**: Full MCP server with schema generation
- **Function Registry**: Proper MCP function definitions with typing
- **Multi-Agent Communication**: Structured data flow between agents
- **Error Handling**: Robust error management across agent interactions
## πŸ“Š Performance
- **Response Times**: Optimized for sub-second agent responses
- **Scalability**: Handles concurrent requests efficiently
- **Reliability**: Built-in fault tolerance and monitoring
- **Resource Management**: Intelligent caching and pooling
## πŸ” Technical Details
- **Python**: 3.12+ required
- **Framework**: Gradio with MCP server capabilities
- **Execution**: Modal for secure sandboxed code execution
- **Search**: Tavily API for real-time web research
- **Monitoring**: Comprehensive performance and health tracking
---
**Ready to experience the future of AI-assisted research and development?**
Start by configuring your API keys and dive into the world of multi-agent AI collaboration! πŸš€