{"id":"llama-mcp-server","name":"llama-mcp-server","af_score":null,"security_score":null,"reliability_score":null,"what_it_does":"Local LLM inference MCP server powered by llama.cpp","best_when":null,"avoid_when":null,"last_evaluated":null}