trt-llm-rag-windows icon indicating copy to clipboard operation
trt-llm-rag-windows copied to clipboard

Installed both models in Chat-With-RTX but can only use Mistral 7B int4

Open Jason-XII opened this issue 1 year ago • 5 comments

I followed the installation instructions and the installer reported that I have successfully installed both models - Llama and Mistral. However when I started the app-launcher.bat file, it said:

Environment path found: C:\Users\Jason\AppData\Local\NVIDIA\ChatWithRTX\env_nvd_rag
App running with config
 {
    "models": {
        "supported": [
            {
                "name": "Mistral 7B int4",
                "installed": true,
                "metadata": {
                    "model_path": "model\\mistral\\mistral7b_int4_engine",
                    "engine": "llama_float16_tp1_rank0.engine",
                    "tokenizer_path": "model\\mistral\\mistral7b_hf",
                    "max_new_tokens": 1024,
                    "max_input_token": 7168,
                    "temperature": 0.1
                }
            },
            {
                "name": "Llama 2 13B int4",
                "installed": false,
                "metadata": {
                    "model_path": "model\\llama\\llama13_int4_engine",
                    "engine": "llama_float16_tp1_rank0.engine",
                    "tokenizer_path": "model\\llama\\llama13_hf",
                    "max_new_tokens": 1024,
                    "max_input_token": 3900,
                    "temperature": 0.1
                }
            }
        ],
        "selected": "Mistral 7B int4"
    },
...

It said that Llama is not installed, but it should be! What might be the possible cause of this problem? How can I fix it? Should I follow the README file instead? Extra information: I am a China user, using an (almost) new computer with a RTX 3060 GPU and 16GB of RAM.

Jason-XII avatar Feb 16 '24 01:02 Jason-XII

看上去这是正常的,至少我展示和你差不多

hbwhzk avatar Feb 16 '24 10:02 hbwhzk

@hbwhzk 你也只能用一个模型吗?我听说使用Llama模型需要十几个G的内存,也许是我16G内存太小了,所以不能用。

Jason-XII avatar Feb 16 '24 11:02 Jason-XII

@Jason-XII 可以用

hbwhzk avatar Feb 18 '24 02:02 hbwhzk

如果确认文件路径正确可以考虑尝试直接修改配置文件,在官网提供的安装程序中是有一个verify_install.py文件来检查模型文件是否正确安装并修改配置文件的

noahc1510 avatar Feb 27 '24 04:02 noahc1510

On which GPU you are trying? On 8GB system only mistral 7B will work. For lama 13B you need GPU with 16GB vmem

anujj avatar May 23 '24 10:05 anujj