mlc-llm icon indicating copy to clipboard operation
mlc-llm copied to clipboard

Universal LLM Deployment Engine with ML Compilation

Results 578 mlc-llm issues
Sort by recently updated
recently updated
newest added

## 🚀 Feature Similar to a standard request to the API, the final stream chunk should have a usage field to give prompt/generated token counts. This can be sent with...

feature request

## 🐛 Bug ## To Reproduce Steps to reproduce the behavior: 1. Build from source using MLC's TVM-Relax & their MLC-LLM Github Repos. I used a few custom options if...

bug

## 🚀 Feature Embeddings support for iOS. Is this part of the roadmap? ## Motivation Ability to generate embeddings in iOS applications. ## Additional context Looking at the code I...

feature request

## 🚀 Feature Add a field to the `ModelRecord` to indicate whether a model is a vision model like `Phi-vision` models. ``` data class ModelRecord( @SerializedName("model_url") val modelUrl: String, @SerializedName("model_id")...

feature request

## 🐛 Bug MLCengine code in quickstart guide on CPU fails with > 'InternalError: Check failed: (it != n->end()) is false: cannot find the corresponding key in the Map' followed...

bug

This PR adds the support for vllm and llama.cpp backend, especially for json generation.

I am trying to use the direct output of `MLCEngine` class but I do not have any clue how to get the output based on the MLC-LLM docs. Is there...

question

## 🐛 Bug The service started based on Meta-Llama-3.1-70B-Instruct fp8 will crash when running a large concurrency. ## To Reproduce ### convert model refer this issue: #2982 ### start service...

bug