{"id":"reprompt-dev-reprompt","name":"reprompt","homepage":"https://getreprompt.dev","repo_url":"https://github.com/reprompt-dev/reprompt","category":"devtools","subcategories":[],"tags":["ai","prompt-engineering","cli","privacy","local-first","mcp","analytics","devtools","python"],"what_it_does":"reprompt-cli (re:prompt) is a local-first CLI tool to scan and analyze prompts generated in AI coding tools. It provides a prompt dashboard, prompt scoring (research-calibrated), conversation distillation/compression, privacy checks for sensitive content exposure, and optional integrations such as a browser extension and an MCP server for certain tools. It claims analysis runs locally with no LLM calls/network requests for the scoring flows.","use_cases":["Score and benchmark prompts for quality (structure/context/position/repetition/clarity)","Distill important turns from long debugging/feature-development conversations","Compress prompts to reduce token usage while preserving intent","Scan local machine/tool artifacts to discover prompts automatically (e.g., Claude Code/Cursor/Cline/Aider/Codex/Cursor)","Run a prompt quality linter in CI (e.g., score thresholds, PR comments)","Privacy review to understand what prompt/data was sent to which AI tools","Agent workflow analysis to detect error loops/tool patterns and identify inefficient sessions"],"not_for":["A network-based hosted service for prompt optimization (it’s primarily local CLI tooling)","Regulated/guaranteed compliance workflows without independently verifying data handling and telemetry behavior","Using as an LLM inference provider or general chatbot","Environments requiring a documented REST/GraphQL API contract (none is evidenced in provided materials)"],"best_when":"You want fast, local, repeatable prompt analytics across multiple coding assistants and you value privacy (local analysis; optional anonymous feature-vector telemetry).","avoid_when":"You need a fully documented programmatic web API (OpenAPI/SDK) or strict, guaranteed no-telemetry guarantees without further audit.","alternatives":["Promptfoo (prompt testing/evals)","LangSmith/Langfuse (hosted tracing/evals)","OpenAI Evals / custom evaluation harnesses","Local prompt scoring libraries (custom scripts using rule-based heuristics)","Microsoft/LLMOps eval frameworks (e.g., Ragas-style for LLM tasks, if adapted)"],"af_score":52.8,"security_score":53.5,"reliability_score":38.8,"package_type":"mcp_server","discovery_source":["github"],"priority":"high","status":"evaluated","version_evaluated":null,"last_evaluated":"2026-03-30T15:25:38.814541+00:00","interface":{"has_rest_api":false,"has_graphql":false,"has_grpc":false,"has_mcp_server":true,"mcp_server_url":null,"has_sdk":false,"sdk_languages":[],"openapi_spec_url":null,"webhooks":false},"auth":{"methods":["None for core local analysis (implied)"],"oauth":false,"scopes":false,"notes":"No authentication scheme described for the CLI. MCP integration is mentioned as an optional extra, but auth configuration for MCP is not documented in the provided materials."},"pricing":{"model":null,"free_tier_exists":false,"free_tier_limits":null,"paid_tiers":[],"requires_credit_card":false,"estimated_workload_costs":null,"notes":"No pricing model described in provided materials; appears to be a local CLI package. Optional telemetry and optional integrations (e.g., Ollama/OpenAI) may incur downstream costs depending on user configuration."},"requirements":{"requires_signup":false,"requires_credit_card":false,"domain_verification":false,"data_residency":[],"compliance":[],"min_contract":null},"agent_readiness":{"af_score":52.8,"security_score":53.5,"reliability_score":38.8,"mcp_server_quality":45.0,"documentation_accuracy":70.0,"error_message_quality":0.0,"error_message_notes":null,"auth_complexity":100.0,"rate_limit_clarity":0.0,"tls_enforcement":50.0,"auth_strength":60.0,"scope_granularity":30.0,"dependency_hygiene":70.0,"secret_handling":60.0,"security_notes":"Claims local analysis with no prompt text leaving the machine for core scoring. Includes privacy tooling and optional telemetry that sends only anonymous feature vectors (not prompt text) per README. However, no concrete details are provided here about telemetry transport/security, MCP auth, or how sensitive data is handled in logs/config files. Optional integrations (ollama/openai) could involve network calls and user-supplied API keys, so secret handling depends on implementation beyond the provided excerpts.","uptime_documented":0.0,"version_stability":70.0,"breaking_changes_history":40.0,"error_recovery":45.0,"idempotency_support":"false","idempotency_notes":"Not specified. CLI actions like scoring/distillation are likely read-only on local data, but behavior for imports/exports is not detailed here.","pagination_style":"none","retry_guidance_documented":false,"known_agent_gotchas":["No evidence in provided materials of an agent-oriented stable API/contract for MCP tool schemas, pagination, or retry semantics.","Most functionality appears local; agents expecting network-based behavior may need file-system/tool artifact access (e.g., scanning installed tools) that depends on local environment state."]}}