jan icon indicating copy to clipboard operation
jan copied to clipboard

bug: Failed! Model Meta-Llama-3.1-8B-Instruct-128k-Q4_0 failed to start.

Open SimplyProgrammer opened this issue 1 year ago • 4 comments

  • [X] I have searched the existing issues

Current behavior

I am unable to start the locally imported model from gpt4all based on Llama-3.1. It simply will not start, sometimes it will display "Failed! Model Meta-Llama-3.1-8B-Instruct-128k-Q4_0 failed to start." in the top corner but even that is not consistent so it is hard to tell what is happening...

Minimum reproduction step

  1. Download Meta-Llama-3.1-8B-Instruct-128k-Q4_0.gguf
  • In my case, I have downloaded it using gpt4all and then in Jan, I have imported it locally and checked the option to copy it to Jan directory. (I tried both options but neither of them worked...)
  1. Try to chat with it -> Nothing happens. Or try to activate it -> it fails...

Expected behavior

The model should start and I should be able to chat with it, or at least get some sort of more consistent error prompts so at least I know that something has happened even if it failed...

Screenshots / Logs

app.log

Jan version

0.5.2

In which operating systems have you tested?

  • [ ] macOS
  • [X] Windows
  • [ ] Linux

Environment details

Windows 10 Intel Core I9 RTX 3080 ti (recently updated from 2080 but it didn't work on it either...) 64GB ram

SimplyProgrammer avatar Aug 03 '24 21:08 SimplyProgrammer

h‌i there, we are working on a huge refa‌c‌tor of Jan, ‌would you mi‌nd trying again with Jan x Cortex via nightly bui‌‌ld? https://github.com/janhq/jan?tab=readme-ov-file#download Thank you

Van-QA avatar Aug 05 '24 06:08 Van-QA

Hi, thanks for the answer. I have tried the nightly build 0.5.2-573 and it indeed works as it should in this version! But unfortunately, this version does not seem to support any kind of hardware (gpu) acceleration and therefore the speed of response generation is, let's say questionable... But I understand that this is nightly build tho, hope it will make it to the full release soon.

SimplyProgrammer avatar Aug 05 '24 19:08 SimplyProgrammer

I have the same issue, looking forward to it in the full release.

anthonyalayo avatar Aug 06 '24 00:08 anthonyalayo

For what it's worth, I tried this out on the nightly and it's failing in the same way: https://huggingface.co/bartowski/Meta-Llama-3.1-8B-Instruct-GGUF

anthonyalayo avatar Aug 06 '24 04:08 anthonyalayo

Hi there, sorry for the inconvenience, due to a roll back in Jan nightly, we are in progress of making Jan support Llama3.1 again, we will keep you posted in any news 🙏

Van-QA avatar Aug 21 '24 07:08 Van-QA

cortex-cpp bump (0.5.0) should address this.

louis-jan avatar Aug 28 '24 03:08 louis-jan

Image

tested and looking good uisng Jan v0.5.2-619

Van-QA avatar Aug 29 '24 04:08 Van-QA

Jan 0.5.3 is released to stable version of Jan app: https://github.com/janhq/jan/releases/tag/v0.5.3, feel free to get back to us if any concern

Van-QA avatar Aug 30 '24 02:08 Van-QA

Well, in my case, the problem has wery much persisted also in 0.5.3 :/ I have tried both importing the GPT4ALLs version of the model and also from Jan model browser under name "Llama 3.1 8B !4 Ins" but in both cases it will not start... Its worth mentioning that I have also Llama 3.0 installedm could this be a problem?

SimplyProgrammer avatar Sep 05 '24 08:09 SimplyProgrammer

I am sorry but it seems like my previous message was not entirely correct. The problem seems to be in the "cortex" API being sneakily uninstalled by the 0.5.3 update. After manually installing it back, it indeed works so my apologies, great job!

SimplyProgrammer avatar Sep 07 '24 20:09 SimplyProgrammer

Hi, I have issue too when loading the model "Meta-Llama-3.1-8B-Instruct-128k-Q4_0" In parameters, if I select my both graphic cards, the model is not loading, but in selecting only one, the model is loaded, but still slow (0.76t/s) but the graphic card seems to be used

thonore75 avatar Sep 15 '24 17:09 thonore75