mlx-omni-server

MLX Omni Server is a local inference server powered by Apple's MLX framework, specifically designed for Apple Silicon (M-series) chips. It implements OpenAI-compatible API endpoints, enabling seamless integration with existing OpenAI SDK clients while leveraging the power of local ML inference.

Python v0.5.2 πŸ“¦ 2.0K/mo
674 stars  Β·  83 forks  Β·  Updated  Β·  MIT
Install
pip install mlx-omni-server
View on GitHub
πŸ“„ Documentation
function-callinggenaimlxopenaiopenai-apistructured-outputstttoolstts
Version history All releases β†—

## What's Changed * Logger now obeys `--log-level` command line arg. Added CORS support by @aperaham in https://github.com/madroidmaq/mlx-omni-server…

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.5.0...v0.5.1

## What's Changed * Support Anthropic API by @madroidmaq in https://github.com/madroidmaq/mlx-omni-server/pull/66 **Full Changelog**: https://gi…

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.4.8...v0.4.9

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.4.6...v0.4.8

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.4.5...v0.4.6

**Full Changelog**: https://github.com/madroidmaq/mlx-omni-server/compare/v0.4.4...v0.4.5

## What's Changed * Support speculative-decoding/draft-model by @madroidmaq in https://github.com/madroidmaq/mlx-omni-server/pull/52 * Refine model …

Quality signals
Score
Stars674
Forks83
Last updated
LicenseMIT
Supported APIs (2)
apple.comapple.com
πŸ“¦ Package statsPyPI β†—
Monthly downloads2.0K