⚡ Native MLX Swift LLM inference server for Apple Silicon. OpenAI-compatible API, SSD streaming for 100B+ MoE models, TurboQuant KV cache compression, + iOS iPhone app.
Stars
196
Forks
6
Watchers
196
Open Issues
1
Overall repository health assessment
No package.json found
This might not be a Node.js project
101
commits
chore: Update Gemma 4 benchmark metrics and add comprehensive testing suite
5a05548View on GitHubMerge branch 'feature/unified-chat-ui' into feature/swiftbuddy
1e3ba78View on GitHubdocs: remove duplicate GIF embed, keep single intro line for iOS 13 Pro 6GB
049bce7View on GitHubdocs: add iPhone 13 Pro 6GB live demo GIF intro to iOS section
cf19434View on GitHubfix(ci): compile default.metallib from .metal sources instead of searching for binary
afb677cView on GitHubdocs: warn against Python mlx-metal metallib version mismatch
33e1511View on GitHubfix(release): correct metallib source — it ships in mlx-swift submodule, not built by swift build
e6556fcView on GitHubdocs: add flash-moe reference to README and introduce benchmark test script
11e7078View on GitHubchore: bump mlx-swift-lm submodule (iOS I/O fix, ExpertStreaming, Mistral4)
1922374View on GitHubdocs: add iOS demo GIF, iOS build instructions, and contributor Team ID note
c22abd0View on GitHub