llama-stack
llama-stack copied to clipboard
Composable building blocks to build Llama Apps
### System Info ``` PyTorch version: 2.5.1+cu124 Is debug build: False CUDA used to build PyTorch: 12.4 ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.5 LTS (x86_64) GCC version:...
### 🚀 Describe the new functionality needed The GET /models/list API should include a new field, tags, in its response. The tags field will provide a way to categorize models...
# What does this PR do? In short, provide a summary of what this PR does and why. Usually, the relevant context should be present in a linked issue. -...
# What does this PR do? This PR adds SambaNova as one of the Provider - Add SambaNova as a provider ## Test Plan Test the functional command ``` pytest...
### System Info :128: RuntimeWarning: 'torch.utils.collect_env' found in sys.modules after import of package 'torch.utils', but prior to execution of 'torch.utils.collect_env'; this may result in unpredictable behaviour Collecting environment information... PyTorch...
# What does this PR do? adds nvidia template for creating a distribution using inference adapter for NVIDIA NIMs. ## Test Plan Please describe: Build llama stack distribution for nvidia...
# What does this PR do? The assert errors are not returned to the client and hid by the 500 internal server error. Append the real error in the error...
This PR adds support of non-cuda XPU backend device into meta-reference stack path. Submitting as a draft PR to facilitate discussion around another patch in llama-models: * https://github.com/meta-llama/llama-models/pull/233 Requires: https://github.com/meta-llama/llama-models/pull/233
# What does this PR do? This PR adds a [SambaNova](https://cloud.sambanova.ai/) inference adapter that enables integration with SambaNova's AI models through their OpenAI-compatible API. Key features implemented: - Chat completion...
llama-stack install from source:https://github.com/meta-llama/llama-stack/tree/cherrypick-working ### System Info python -m "torch.utils.collect_env" /home/kaiwu/miniconda3/envs/llama/lib/python3.10/runpy.py:126: RuntimeWarning: 'torch.utils.collect_env' found in sys.modules after import of package 'torch.utils', but prior to execution of 'torch.utils.collect_env'; this may result...