vllm-mlx

OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP tool calling, and multimodal support. Native MLX backend, 400+ tok/s. Works with Claude Code.

Repo ↗ AI & Machine Learning anthropic apple-silicon audio-processing claude-code computer-vision image-understanding inference llm machine-learning macos mllm mlx multimodal-ai speech-to-text stt text-to-speech tts video-understanding vision-language-model vllm
⚙ Agent Friendliness
N/A
Not evaluated
Can an agent use this?
🔒 Security
N/A
Not evaluated
Is it safe for agents?
⚡ Reliability
N/A
Not evaluated
Does it work consistently?

Scores are editorial opinions as of unknown date.

5186
Packages Evaluated
26151
Need Evaluation
173
Need Re-evaluation
Community Powered