IntelliServer
IntelliServer copied to clipboard
Create offline model loader, to server llama model from local repository
Add a service to load llama models family and serve them through an API, the official python code to server the models available here:
https://github.com/facebookresearch/llama
Request the models access: https://ai.meta.com/resources/models-and-libraries/llama-downloads/
LLama 2 Onix, which is more suitable to server in node js setup:
https://github.com/microsoft/Llama-2-Onnx