{"id":"disler-just-prompt","name":"just-prompt","homepage":null,"repo_url":"https://github.com/disler/just-prompt","category":"ai-ml","subcategories":[],"tags":["mcp","llm","routing","multi-provider","tooling","prompting"],"what_it_does":"just-prompt is an MCP (Model Control Protocol) server that exposes a unified interface to multiple LLM providers (OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, and Ollama). It provides MCP tools to send prompts to one or more models, run prompts loaded from files (optionally writing outputs to disk), list providers/models, and run a multi-model “board” workflow where a “CEO” model decides based on board member responses.","use_cases":["Unified prompt execution across multiple LLM providers","Benchmarking or fallback across multiple models/providers","Generating outputs from prompt templates stored in files","Running a multi-model consensus/selection workflow (board/CEO pattern)","Local and remote model usage (including Ollama via a host URL)"],"not_for":["Producing an internet-accessible hosted API for external clients (it’s an MCP server intended to be run locally/in-process)","Use cases requiring strict data residency or compliance guarantees (not documented here)","Use cases needing fine-grained API-level RBAC/tenant security (not documented here)"],"best_when":"You want a local MCP server that an agent can call to route prompts across multiple LLM backends with minimal integration effort.","avoid_when":"You need strong organizational security controls (authZ, audit logs), guaranteed idempotency for file-writing operations, or well-specified operational/SLA guarantees.","alternatives":["Direct provider SDKs/APIs (OpenAI/Anthropic/Gemini/etc.)","Other MCP servers or LLM routers (e.g., provider-specific MCP implementations, or general LLM gateway services)"],"af_score":62.0,"security_score":47.5,"reliability_score":21.2,"package_type":"mcp_server","discovery_source":["github"],"priority":"high","status":"evaluated","version_evaluated":null,"last_evaluated":"2026-03-30T13:25:50.052637+00:00","interface":{"has_rest_api":false,"has_graphql":false,"has_grpc":false,"has_mcp_server":true,"mcp_server_url":null,"has_sdk":false,"sdk_languages":[],"openapi_spec_url":null,"webhooks":false},"auth":{"methods":["Environment-variable API keys for supported providers (OPENAI_API_KEY, ANTHROPIC_API_KEY, GEMINI_API_KEY, GROQ_API_KEY, DEEPSEEK_API_KEY)","Ollama connection via OLLAMA_HOST (no explicit auth described)"],"oauth":false,"scopes":false,"notes":"Authentication/authorization is described as local API-key configuration via environment variables. The MCP interface itself does not document additional auth, scopes, or multi-tenant access controls."},"pricing":{"model":null,"free_tier_exists":false,"free_tier_limits":null,"paid_tiers":[],"requires_credit_card":false,"estimated_workload_costs":null,"notes":"Pricing is not provided by just-prompt; costs depend on the underlying provider usage (and local Ollama is typically self-hosted)."},"requirements":{"requires_signup":false,"requires_credit_card":false,"domain_verification":false,"data_residency":[],"compliance":[],"min_contract":null},"agent_readiness":{"af_score":62.0,"security_score":47.5,"reliability_score":21.2,"mcp_server_quality":88.0,"documentation_accuracy":75.0,"error_message_quality":0.0,"error_message_notes":null,"auth_complexity":80.0,"rate_limit_clarity":20.0,"tls_enforcement":60.0,"auth_strength":45.0,"scope_granularity":10.0,"dependency_hygiene":55.0,"secret_handling":70.0,"security_notes":"TLS is not discussed for MCP/stdio or provider-to-provider calls. Auth is based on environment-provided provider API keys with no documented MCP-level authorization or scopes; this limits isolation for multi-user environments. Secret handling is only implied by using environment variables; there is no explicit guidance about logging/redaction. Dependency hygiene is assumed but not verifiable from the provided content (dependencies listed; no vulnerability/SBOM/CVE info provided).","uptime_documented":0.0,"version_stability":35.0,"breaking_changes_history":20.0,"error_recovery":30.0,"idempotency_support":"false","idempotency_notes":"Some tools write output files (prompt_from_file_to_file and ceo_and_board). Idempotency is not stated; repeated calls could overwrite or duplicate outputs depending on implementation defaults (not described here).","pagination_style":"none","retry_guidance_documented":false,"known_agent_gotchas":["File path parameters require absolute paths (abs_file_path/abs_output_dir); agents should ensure host filesystem access matches the server environment.","Model names require provider prefixes (e.g., openai:o3:high). The server mentions automatic model name correction based on default models, but exact behavior is not fully specified.","Provider availability depends on which API keys are present; missing keys mean some providers become unavailable while the server still starts. Agents should handle partial provider sets."]}}