llama-stack icon indicating copy to clipboard operation
llama-stack copied to clipboard

Composable building blocks to build Llama Apps

Results 360 llama-stack issues
Sort by recently updated
recently updated
newest added
trafficstars

### System Info ``` PyTorch version: 2.5.1+cu124 Is debug build: False CUDA used to build PyTorch: 12.4 ROCM used to build PyTorch: N/A OS: Ubuntu 22.04.5 LTS (x86_64) GCC version:...

bug

### 🚀 Describe the new functionality needed The GET /models/list API should include a new field, tags, in its response. The tags field will provide a way to categorize models...

# What does this PR do? In short, provide a summary of what this PR does and why. Usually, the relevant context should be present in a linked issue. -...

CLA Signed

# What does this PR do? This PR adds SambaNova as one of the Provider - Add SambaNova as a provider ## Test Plan Test the functional command ``` pytest...

CLA Signed

### System Info :128: RuntimeWarning: 'torch.utils.collect_env' found in sys.modules after import of package 'torch.utils', but prior to execution of 'torch.utils.collect_env'; this may result in unpredictable behaviour Collecting environment information... PyTorch...

# What does this PR do? adds nvidia template for creating a distribution using inference adapter for NVIDIA NIMs. ## Test Plan Please describe: Build llama stack distribution for nvidia...

CLA Signed

# What does this PR do? The assert errors are not returned to the client and hid by the 500 internal server error. Append the real error in the error...

CLA Signed

This PR adds support of non-cuda XPU backend device into meta-reference stack path. Submitting as a draft PR to facilitate discussion around another patch in llama-models: * https://github.com/meta-llama/llama-models/pull/233 Requires: https://github.com/meta-llama/llama-models/pull/233

CLA Signed

# What does this PR do? This PR adds a [SambaNova](https://cloud.sambanova.ai/) inference adapter that enables integration with SambaNova's AI models through their OpenAI-compatible API. Key features implemented: - Chat completion...

CLA Signed

llama-stack install from source:https://github.com/meta-llama/llama-stack/tree/cherrypick-working ### System Info python -m "torch.utils.collect_env" /home/kaiwu/miniconda3/envs/llama/lib/python3.10/runpy.py:126: RuntimeWarning: 'torch.utils.collect_env' found in sys.modules after import of package 'torch.utils', but prior to execution of 'torch.utils.collect_env'; this may result...

RAG