GPTeam icon indicating copy to clipboard operation
GPTeam copied to clipboard

Local Language Model Implementation Options: Vicuna and Open-Source Alternatives

Open lexis33 opened this issue 1 year ago • 4 comments

It would be fantastic to have the possibility of running it on local Language Model implementations such as Vicuna or other open-source alternatives.

lexis33 avatar Jun 08 '23 10:06 lexis33

I recently set it up to run on my local 7B Vicuna model via a custom Langchain classes for VicunaLLM and VicunaChatModel but the issue is that the responses tend to not return in proper JSON format and the system breaks down. So just letting you know that Vicuna is probably not really good enough to work out of the box without some additional tweaking. I could potentially look into LangChain's ReLLM integration to try and get around this problem though. I'm also looking to tweak my setup to get the 13B Vicuna running with CPU assistance. Just figured I'd share my experience with Vicuna here since I saw your comment.

I'm also brand new to this stuff so I could just have messed something up too.

cstan969 avatar Jun 09 '23 01:06 cstan969

I recently set it up to run on my local 7B Vicuna model via a custom Langchain classes for VicunaLLM and VicunaChatModel but the issue is that the responses tend to not return in proper JSON format and the system breaks down. So just letting you know that Vicuna is probably not really good enough to work out of the box without some additional tweaking. I could potentially look into LangChain's ReLLM integration to try and get around this problem though. I'm also looking to tweak my setup to get the 13B Vicuna running with CPU assistance. Just figured I'd share my experience with Vicuna here since I saw your comment.

I'm also brand new to this stuff so I could just have messed something up too.

Well, Thank you. I would really like to see if you get it run and if you can later share the code with maybe video or tutorial it will be great. Thank you!

lexis33 avatar Jun 09 '23 03:06 lexis33

Generally I'd say the formatting guidance in the prompts is not strong enough, since some of the prompts get too long.

I reckon it is possible to get a version running with smaller models like Vicuna 7B (albeit very slowly), if some of the prompts are reduced in length, so this would be a great place to start if you are looking to go that route.

joshsny avatar Jun 09 '23 07:06 joshsny

Generally I'd say the formatting guidance in the prompts is not strong enough, since some of the prompts get too long.

I reckon it is possible to get a version running with smaller models like Vicuna 7B (albeit very slowly), if some of the prompts are reduced in length, so this would be a great place to start if you are looking to go that route.

Thanks for the information.

Can you tell how you did it? It's asking for API from open ai. I have little experience with python. I overall use oobabooga mostly.

lexis33 avatar Jun 09 '23 23:06 lexis33