{"id":"mcpstore","name":"MCPStore","homepage":"https://github.com/whillhill/mcpstore","repo_url":"https://github.com/whillhill/mcpstore","category":"sdk","subcategories":["python-library","orchestration","agent-framework-adapter"],"tags":["python","langchain","langgraph","autogen","crewai","llamaindex","redis","fastapi","orchestration","pip"],"what_it_does":"MCPStore is a Python orchestration layer that manages MCP service lifecycles and adapts them for use with popular AI agent frameworks like LangChain, LangGraph, AutoGen, CrewAI, and LlamaIndex through a unified API.","use_cases":["Building LangChain or LangGraph agents that consume multiple MCP tools without manual protocol management","Isolating MCP tool sets per agent to prevent context overflow in multi-agent systems","Running a REST API server that exposes MCP service management endpoints for external orchestration"],"not_for":["Users who need a standalone MCP gateway deployable without Python code","Teams that do not use LangChain, AutoGen, CrewAI, or similar Python agent frameworks","Production deployments requiring enterprise auth, RBAC, or SLA guarantees"],"best_when":"You are building Python-based AI agents with LangChain or similar frameworks and want managed MCP service connections with hot-swap and per-agent tool isolation.","avoid_when":"You need language-agnostic gateway functionality, or your agent stack is outside Python.","alternatives":["mcpjungle","metamcp","mcpo"],"af_score":58.4,"security_score":45.0,"reliability_score":null,"package_type":"mcp_server","discovery_source":["github"],"priority":"low","status":"evaluated","version_evaluated":"latest","last_evaluated":"2026-03-01T09:50:05.931314+00:00","performance":{"latency_p50_ms":null,"latency_p99_ms":null,"uptime_sla_percent":null,"rate_limits":null,"data_source":"llm_estimated","measured_on":null}}