{"id":"ai-assistant-core","name":"ai_assistant_core","homepage":"https://ai-assistant.runawaybrains.com","repo_url":"https://github.com/OrlandoLuque/ai_assistant_core","category":"ai-ml","subcategories":[],"tags":[],"what_it_does":"Simple, ergonomic Rust client & server for local LLMs (Ollama, LM Studio, OpenAI-compatible). Chat, list models, stream responses, serve your model remotely.","use_cases":[],"not_for":[],"best_when":null,"avoid_when":null,"alternatives":[],"af_score":null,"security_score":null,"reliability_score":null,"package_type":"mcp_server","discovery_source":["crates_io"],"priority":"medium","status":"discovered","version_evaluated":null,"last_evaluated":null}