vllm-mlx
OpenAI and Anthropic compatible server for Apple Silicon. Run LLMs and vision-language models (Llama, Qwen-VL, LLaVA) with continuous batching, MCP tool calling, and multimodal support. Native MLX backend, 400+ tok/s. Works with Claude Code.
Repo ↗
AI & Machine Learning
anthropic
apple-silicon
audio-processing
claude-code
computer-vision
image-understanding
inference
llm
machine-learning
macos
mllm
mlx
multimodal-ai
speech-to-text
stt
text-to-speech
tts
video-understanding
vision-language-model
vllm
⚙ Agent Friendliness
N/A
Not evaluated
Can an agent use this?
🔒 Security
N/A
Not evaluated
Is it safe for agents?
⚡ Reliability
N/A
Not evaluated
Does it work consistently?
Scores are editorial opinions as of unknown date.