OnnxStack
OnnxStack copied to clipboard
TextGeneration Pipeline
Support for TextGeneration ONNX models
Initial support will be using OnnxRuntime-GenAI onnxruntime-genai
TODO:
- [x] TextGeneration Project
- [ ] Basic stateless pipeline
- [ ] CUDA and CPU support
- [ ] Code Examples
Looks cool. How is the text generation speed for onnx models compared to llamasharp for a Mistral 7B v0.2 Instruct model for example?
Looks cool. How is the text generation speed for onnx models compared to llamasharp for a Mistral 7B v0.2 Instruct model for example?
Im not too sure, I have not been able to get LLamaSharp working in Web or WPF since version 0.8. so this is my first attempt using another LLM base library, this one is still very new but its a million times less complicated than llama.cpp