mlc-llm
mlc-llm copied to clipboard
What is the best way to deploy the 'mlc_chat' service as an API that accepts HTTP requests?
I tested the 'mlc_chat_cli' command in a Linux environment, and I want to deploy it as an API that can call services. Is there a convenient way to deploy it, such as a Docker image?
Not currently, but we are in the process of putting up a PR soon that will expose a REST API.
Please check out https://github.com/mlc-ai/mlc-llm/tree/main/python#rest-api for instructions on how to use the REST API.
closing as rest api is now part of the main