{"id":"ai-assistant-core","name":"ai_assistant_core","af_score":null,"security_score":null,"reliability_score":null,"what_it_does":"Simple, ergonomic Rust client & server for local LLMs (Ollama, LM Studio, OpenAI-compatible). Chat, list models, stream responses, serve your model remotely.","best_when":null,"avoid_when":null,"last_evaluated":null}