{"id":"houtini-lm","name":"@houtini/lm","homepage":"https://houtini.ai","repo_url":"git+https://github.com/houtini-ai/lm.git","category":"ai-ml","subcategories":[],"tags":["mcp","model-context-protocol","mcp-server","lm-studio","ollama","vllm","openai","openai-compatible","local-llm","claude","ai-tools","llama-cpp","ai","llm"],"what_it_does":"MCP server for local LLMs — connects to LM Studio or any OpenAI-compatible endpoint","use_cases":[],"not_for":[],"best_when":null,"avoid_when":null,"alternatives":[],"af_score":null,"security_score":null,"reliability_score":null,"package_type":"mcp_server","discovery_source":["npm"],"priority":"medium","status":"discovered","version_evaluated":null,"last_evaluated":null}