{"id":"llm-wrapper-mcp-server","name":"llm-wrapper-mcp-server","homepage":"https://pypi.org/project/llm-wrapper-mcp-server/","repo_url":null,"category":"ai-ml","subcategories":[],"tags":[],"what_it_does":"Wrap a call to any remote LLM model and expose it as an MCP server tool to allow your main model to communicate with other models.","use_cases":[],"not_for":[],"best_when":null,"avoid_when":null,"alternatives":[],"af_score":null,"security_score":null,"reliability_score":null,"package_type":"mcp_server","discovery_source":["pypi"],"priority":"low","status":"discovered","version_evaluated":null,"last_evaluated":null}