Skip to content

Add Rapid-MLX to Large Model Serving#310

Open
raullenchai wants to merge 1 commit intotensorchord:mainfrom
raullenchai:add-rapid-mlx
Open

Add Rapid-MLX to Large Model Serving#310
raullenchai wants to merge 1 commit intotensorchord:mainfrom
raullenchai:add-rapid-mlx

Conversation

@raullenchai
Copy link
Copy Markdown

Summary

  • Adds Rapid-MLX to the Large Model Serving section

Rapid-MLX is an OpenAI-compatible LLM inference server optimized for Apple Silicon using MLX. It provides 2-4x faster inference than Ollama, with full tool calling support, reasoning separation, and prompt caching. Apache-2.0 licensed.

nicoloboschi pushed a commit to nicoloboschi/Awesome-LLMOps that referenced this pull request Mar 30, 2026
…onary Framework category (tensorchord#310)

Co-authored-by: kerthcet <kerthcet@users.noreply.github.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant