{"id":"corbell-ai-corbell","name":"Corbell","af_score":59.5,"security_score":44.8,"reliability_score":21.2,"what_it_does":"Corbell is a local CLI tool (Python) that builds multi-repo architecture graphs from code (SQLite by default), indexes code embeddings, extracts design patterns from existing docs, and uses LLM providers to generate and review PRD-driven architecture/spec documents. It can also expose graph/code context via an MCP server and export tasks to Linear (and optionally Notion).","best_when":"You have multiple backend repositories with established architecture patterns and you want repeatable, context-aware spec generation/review entirely from local code.","avoid_when":"You require a strict, audited compliance workflow for AI outputs, or you cannot provide/handle repository source code locally.","last_evaluated":"2026-03-30T13:42:37.542454+00:00","has_mcp":true,"has_api":false,"auth_methods":["Environment variables for provider API keys (e.g., ANTHROPIC_API_KEY, OPENAI_API_KEY, BEDROCK_API_KEY, AZURE_OPENAI_API_KEY, GOOGLE_APPLICATION_CREDENTIALS)","Optional use of local models via Ollama (no API key required for that provider)"],"has_free_tier":false,"known_gotchas":["Local-first tool: agents must run in a filesystem context with the repositories accessible.","MCP tool behavior depends on the workspace state (graph built/indexed) and environment variables for the LLM provider when LLM-dependent tools are invoked.","Generated specs may require human approval; graph consistency/review steps are part of the intended workflow."],"error_quality":0.0}