{"id":"disler-just-prompt","name":"just-prompt","af_score":62.0,"security_score":47.5,"reliability_score":21.2,"what_it_does":"just-prompt is an MCP (Model Control Protocol) server that exposes a unified interface to multiple LLM providers (OpenAI, Anthropic, Google Gemini, Groq, DeepSeek, and Ollama). It provides MCP tools to send prompts to one or more models, run prompts loaded from files (optionally writing outputs to disk), list providers/models, and run a multi-model “board” workflow where a “CEO” model decides based on board member responses.","best_when":"You want a local MCP server that an agent can call to route prompts across multiple LLM backends with minimal integration effort.","avoid_when":"You need strong organizational security controls (authZ, audit logs), guaranteed idempotency for file-writing operations, or well-specified operational/SLA guarantees.","last_evaluated":"2026-03-30T13:25:50.052637+00:00","has_mcp":true,"has_api":false,"auth_methods":["Environment-variable API keys for supported providers (OPENAI_API_KEY, ANTHROPIC_API_KEY, GEMINI_API_KEY, GROQ_API_KEY, DEEPSEEK_API_KEY)","Ollama connection via OLLAMA_HOST (no explicit auth described)"],"has_free_tier":false,"known_gotchas":["File path parameters require absolute paths (abs_file_path/abs_output_dir); agents should ensure host filesystem access matches the server environment.","Model names require provider prefixes (e.g., openai:o3:high). The server mentions automatic model name correction based on default models, but exact behavior is not fully specified.","Provider availability depends on which API keys are present; missing keys mean some providers become unavailable while the server still starts. Agents should handle partial provider sets."],"error_quality":0.0}