PicoMLXServer
PicoMLXServer copied to clipboard
The easiest way to run the fastest MLX-based LLMs locally
Pico MLX Server
Pico MLX Server is the easiest way to get started with Apple's MLX AI framework
Pico MLX Server provides a GUI for MLX Server. MLX server provides an API for local MLX models conform the OpenAI API. This allows you to use most existing OpenAI chat clients with Pico MLX Server.

Highlights
- Start and stop servers quickly via menu bar extra
- Download MLX models from the MLX community on HuggingFace
- Install and setup Python environment, MLX, and MLX Server from within Pico MLX Server.
- Run multiple servers on different ports
- View logs of the servers in separte windows
- Custom link to open your favorite chat client (defaults to Pico AI Assistant)
Supported MLX models
See MLX Community on HuggingFace
Getting Started
To install Pico MLX Server, build the source using Xcode, or download the notarized executable directly from GitHub.
To set up Pico MLX Server, open the app and
- Install and set up Python, pip, MLX and MLX Server and optionally Conda manually
- Use Pico MLX Server's automated Setup (
MLX -> Setup...) - Pico MLX Server uses Conda by default to create a virtual environment for the servers called pico. This (hopefully) will avoid any Python version issues. Conda can be disabled in
Settings

Requirements
- MacOS 14.0 (Sonoma) or later
Create a New Server

- Select
MLX -> Servers -> New Server... - Press
Createto create the default servermlx-community/Nous-Hermes-2-Mistral-7B-DPO-4bit-MLXon port8080 - To use a different model, click on the
vbutton or type in a model manually from the MLX Community on HuggingFace (make sure to use themlx-community/prefix) - Press the
View Logsbuttons to open a window with the server's real-time logs

Use Pico MLX Server with an AI client
-
Point any OpenAI API compatible AI assistant to
http://127.0.0.1:8080(or any other port you used in Pico MLX Server). (Instructions for Pico AI Assistant coming soon) -
Curl:
curl -X GET 'http://127.0.0.1:8080/generate?prompt=write%20me%20a%20poem%20about%the%20ocean&stream=true'
API Endpoints
- Pico MLX Server uses OpenAI's
POST /v1/completionsAPI. See for more information https://platform.openai.com/docs/api-reference/completions/create.
Known Issues
- Pico MLX Server doesn't detect if a port is already in use (use
lsof -i:8080in the terminal to find the PID of the running server) - There is a SwiftUI issue in
New Serverwindow and theServersmenu where the state of servers isn't updated
Roadmap
- Switch from Python to MLX Swift
- Swift-based HTTP server
Related projects
Pico MLX Server is part of a bundle of open source Swift tools for AI engineers. Looking for a server-side Swift OpenAI proxy to protect your OpenAI keys? Check out Swift OpenAI Proxy.
Authors and Acknowledgements
Pico MLX Server, Swift OpenAI Proxy, and Pico AI Assistant were created by Ronald Mannak with help from Ray Fernando
MLX Server was created by Mustafa Aljadery & Siddharth Sharma
Code used from MLX Swift Chat and Swift Chat