llama-stack
llama-stack copied to clipboard
Composable building blocks to build Llama Apps
[2401213419@l12gpu30 llama3.2]$ llama stack run llapku --port 8080 Using config `/lustre/home/2401213419/.llama/builds/conda/llapku-run.yaml` Resolved 19 providers inner-inference => meta-reference models => __routing_table__ inference => __autorouted__ inner-safety => meta-reference shields => __routing_table__ safety...
# What does this PR do? Integrate Portkey AI Inference Provider to Llama Stack. Access to 40+ Providers to access 250+ LLMs with LLama Stack -Fixes and Closes #671 ##...
### 🚀 Describe the new functionality needed Ability to use Portkey AI wth Llama Stack to access opensource models across 40+ providers ### 💡 Why is this needed? What if...
### 🚀 Describe the new functionality needed Something to consider for after our apis are stable and the ecosystem matures a bit - but keen to know your thoughts. We...
# What does this PR do? In short, provide a summary of what this PR does and why. Usually, the relevant context should be present in a linked issue. -...
# What does this PR do? As is. ## Test Plan Doc only ## Sources [repo](https://github.com/bentoml/openllm) ## Before submitting - [x] This PR fixes a typo or improves the docs...
# What does this PR do? - Refactor BaseScoringFn to be just a minimal interface, add new RegistrableBaseScoring - Refactor data schema check - To separately evaluate retrieval component in...
# What does this PR do? - This PR adds the ability s.t. users can evaluate on both retrieval + generation separately & as a whole by passing an AgentConfig...
### 🚀 Describe the new functionality needed We want to look into multi modal retrieval support for llama stack, I want to first discuss what the inference provider API side...
# What does this PR do? - add more braintrust scoring functions for RAG eval - add tests for evaluating against context ## Test Plan ``` pytest -v -s -m...