Automatically crawl, analyze, test, and generate intelligent reports for any developer tool documentation.
DevAgent is a comprehensive testing system that uses AI to evaluate developer tools by analyzing their documentation, generating test cases, executing them, and providing detailed insights for improvement.
DevAgent automates the entire process of evaluating developer tools and APIs:
- Uses Crawl4AI with deep crawling strategies to discover documentation pages
- Supports multiple crawling modes: simple, deep (BFS/DFS), and adaptive crawling
- Smart URL normalization to avoid duplicates (
/apivs/api/) - Filters and focuses on relevant documentation content
- DSPy-powered document analysis that extracts:
- API operations and capabilities
- Authentication methods and requirements
- Usage patterns and workflows
- Error scenarios and edge cases
- Code examples and integration guides
- Generates comprehensive test cases across multiple categories:
- Authentication testing - API key validation, OAuth flows
- Basic usage - Core functionality verification
- Core workflows - Multi-step process testing
- Error handling - Edge cases and failure scenarios
- Prioritizes tests based on complexity and importance
- Runs tests in parallel with configurable worker pools
- Thread-safe execution with isolated contexts
- Real-time progress tracking and error reporting
- Graceful fallback to sequential execution when needed
- AI-generated insights analyzing test failures against documentation
- Page-level reports with specific recommendations
- Overall quality scores and improvement suggestions
- Gap analysis identifying missing examples and unclear documentation
- Web-based dashboard with modern, interactive UI
- ๐ Modern Web Interface - FastAPI-powered dashboard with real-time updates
- ๐ง Flexible Configuration - Customize crawling depth, test parameters, and API keys
- ๐ Progress Tracking - Monitor pipeline execution across all stages
- ๐พ Persistent Results - Save and review past testing runs
- ๐จ Beautiful UI - Modern, responsive design with dark theme
- ๐ Real-time Updates - Auto-refreshing status and progress indicators
- ๐ Comprehensive Logging - Detailed execution traces and error reporting
- Python 3.11+
- uv (recommended) or pip for package management
git clone <repository-url>
cd devagentUsing uv (recommended):
uv syncOr using pip:
pip install -e .uv run playwright installOr if using pip:
playwright installThis downloads the required browser binaries (Chromium, Firefox, WebKit) that Crawl4AI needs for web scraping.
Set up your preferred AI model by setting environment variables:
# For Gemini (recommended)
export GEMINI_API_KEY="your-openai-api-key"
# OR create a .env file
echo "GEMINI_API_KEY=your-key-here" > .envStart the web server:
uv run devagent-webThen open your browser to: http://localhost:8005
The web interface allows you to:
- Configure tool testing parameters
- Set API keys and context variables
- Monitor real-time progress
- View comprehensive reports
- Access historical test runs
For programmatic usage:
uv run devagent-cliOr run directly:
python agents/test.py- Open the web interface at http://localhost:8005
- Enter tool details:
- Tool Name:
OpenWeatherMap API - Base URL:
https://openweathermap.org/api
- Tool Name:
- Add API keys (KEY:VALUE format):
OPENWEATHER_API_KEY:your-api-key-here
- Configure options:
- Max Pages:
20 - Max Depth:
3 - Keywords:
api, documentation, guide
- Max Pages:
- Click "Start Testing Pipeline"
- Monitor progress in real-time
- Review results including:
- Overall quality score
- AI-generated insights
- Page-level analysis
- Specific improvement recommendations
The testing pipeline consists of 5 main stages:
- ๐ท๏ธ Fetching - Crawl and discover documentation pages
- ๐ Analysis - AI-powered content extraction and categorization
- ๐ Test Planning - Generate comprehensive test scenarios
- โก Execution - Run tests in parallel with isolated contexts
- ๐ Reporting - Generate insights and recommendations
Each stage provides detailed progress updates and error handling.
- Max Pages: Maximum number of pages to crawl (1-100)
- Max Depth: How deep to crawl from the base URL (1-5)
- Keywords: Focus keywords for relevance scoring
- URLs to Exclude: Skip specific URLs or patterns
- Max Workers: Number of parallel workers (1-16)
- API Keys: Set testing credentials and context variables
- Timeouts: Configure request and execution timeouts
- Model Selection: Choose between OpenAI GPT or Claude models
- Analysis Depth: Configure how thorough the AI analysis should be
- Modern, responsive design with dark theme
- Dynamic API key management - add/remove key-value pairs
- Advanced options with collapsible sections
- Form validation and user-friendly error messages
- Real-time progress tracking with auto-refresh
- Interactive report viewing with expandable sections
- Search and filtering for large result sets
- Export capabilities for reports and raw data
- Live status updates during execution
- Detailed error reporting with stack traces
- Stage-by-stage progress with timing information
- Background execution without blocking the UI
- Clone and install as described above
- Install development dependencies:
uv sync --dev
- Run tests:
uv run pytest