llama-stack
llama-stack copied to clipboard
Composable building blocks to build Llama Apps
### 🚀 Describe the new functionality needed "stop" parameter: https://platform.openai.com/docs/api-reference/completions/create#completions-create-stop ### 💡 Why is this needed? What if we don't build it? Only vLLM inference provider is supported/tested through https://github.com/meta-llama/llama-stack/pull/1715...
# What does this PR do? [Provide a short summary of what this PR does and why. Link to relevant issues if applicable.] [//]: # (If resolving an issue, uncomment...
Goals: * remove the need of a custom tool to install a collection of python packages AKA `llama stack build` * use the power of 'uv', which was designed to...
### 🚀 Describe the new functionality needed Develop acceptance criteria for llama-stack (MCP with auth) and scope other requirements to enable MCP Servers with both server-side and client-side auth. ###...
# What does this PR do? This PR contains two sets of notebooks that serve as reference material for developers getting started with Llama Stack using the NVIDIA Provider. Developers...
### 🚀 Describe the new functionality needed # Overview The goal for Llama Stack v1 is to enable ISVs and enterprise developers to build AI applications in on-prem and VPC...
# What does this PR do? Prototype on a new feature to allow new APIs to be plugged in Llama Stack. Opened for early feedback on the approach and test...
### 🚀 Describe the new functionality needed The llama stack distro server has several stores for state: 1. kvstore - for routing info 2. sqlstore - to store chat completion...
### 🚀 Describe the new functionality needed Add higher-level APIs, e.g. list chat completions ### 💡 Why is this needed? What if we don't build it? Enable more efficient and...
### 🚀 Describe the new functionality needed Log view for chat completions ### 💡 Why is this needed? What if we don't build it? Implement a log view for chat...