Orion icon indicating copy to clipboard operation
Orion copied to clipboard

Orion-14B is a family of models includes a 14B foundation LLM, and a series of models: a chat model, a long context model, a quantized model, a RAG fine-tuned model, and an Agent fine-tuned model. Ori...

Results 43 Orion issues
Sort by recently updated
recently updated
newest added

llm_load_tensors: ggml ctx size = 0.41 MiB ggml_backend_metal_buffer_from_ptr: error: failed to allocate buffer, size = 0.00 MiB llama_model_load: error loading model: failed to allocate buffer llama_load_model_from_file: failed to load model...

使用 ggfu 自定义 Modelfile, 创建成功,但是 ollama run orion-14b-chat 一直没有响应

启动这个需要配套什么吗? 还有,相关的模型配置哪里?现在似乎看不到配置的地方。

请问插件版本用的哪些数据集呢?是基于base模型加chat版本的SFT数据和插件数据混合训练的吗,还是基于Chat模型再次训练的呢

看介绍说Orion-14B-Chat-Plugin可以支持插件和函数调用,可以出一个简单的例子吗

使用的linux系统,T4卡,安装不上flash-attn库,可以绕过flash-attn库进行运行吗

在[generation_utils.py](https://huggingface.co/OrionStarAI/Orion-14B-Chat/blob/main/generation_utils.py)的`build_chat_input`函数里没有看到system prompt相关的内容,想问一下模型是否支持使用sytem prompt推理/训练?

![WechatIMG16](https://github.com/OrionStarAI/Orion/assets/94165844/77ea1b28-6f96-4e5c-a623-973e89993582) ![WechatIMG17](https://github.com/OrionStarAI/Orion/assets/94165844/0cc5ab5a-01a8-4f2c-a560-2d80a9d96de4) 环境变量如上,但是没有输出结果,对于tool的URL:用的是代码里面自带的,请问下该怎么办

Hi, when i run Orion-14B-Chat-Int4, by following code on A800-80G ``` import torch from transformers import AutoTokenizer, AutoModelForCausalLM, GenerationConfig model_name = "OrionStarAI/Orion-14B-Chat-Int4" tokenizer = AutoTokenizer.from_pretrained(model_name, use_fast=False, trust_remote_code=True) model = AutoModelForCausalLM.from_pretrained(model_name,...

我现在用huggingface 感觉一不小心显存就炸了