{"id":"sadiuysal-crawl4ai-mcp-server","name":"crawl4ai-mcp-server","af_score":64.2,"security_score":44.0,"reliability_score":21.2,"what_it_does":"crawl4ai-mcp-server is a self-hosted MCP server (stdio) that exposes Crawl4AI scraping/crawling capabilities to AI agents via four MCP tools: scrape, crawl, crawl_site, and crawl_sitemap. It supports Markdown extraction, BFS crawling with depth/page limits and optional adaptive stopping, and can optionally persist results to disk via output_dir/manifest files, while applying URL safety blocks for localhost/private/internal targets.","best_when":"Used locally/self-hosted by an agent workflow to gather web content from allowed public URLs and store crawl artifacts for later processing.","avoid_when":"Avoid targeting private/internal networks even if the agent tries; also avoid using it as a generic open proxy without strict egress controls and operational guardrails.","last_evaluated":"2026-03-30T13:53:24.660317+00:00","has_mcp":true,"has_api":false,"auth_methods":["OPENAI_API_KEY environment variable (only referenced for OpenAI Agents SDK example)"],"has_free_tier":false,"known_gotchas":["MCP server is stdio-based; requires correct MCP client setup (command/args) and editor configuration.","Long crawls may be constrained by max_pages/max_depth/timeout_sec; agents should tune these to avoid partial results.","Use of output_dir persists artifacts; agents should manage filesystem/volume permissions and cleanup.","URL safety blocks may prevent scraping expected targets if they match localhost/private/internal patterns."],"error_quality":0.0}