deadpipe
deadpipe
hi am also facing the same problem have you found a solution? @gapjialin
https://github.com/haotian-liu/LLaVA/issues/1230#issuecomment-2014765471 Temporary Fix
Facing same issue here. Have you found a solution, @mkloppel ?
@Maknee I tried setting option(MINIGPT4_CUBLAS "minigpt4: use cuBLAS" ON) in the [CMakeLists.txt](https://github.com/Maknee/minigpt4.cpp/blob/master/CMakeLists.txt#L35). But when i run cmake --build . --config Release, i get this error below unfortunately : - ![C__Windows_System32_cmd...
@Maknee Do you have this on your roadmap to fix this CuBlas integration with minigpt4.cpp?
> same problem @Z500-RAY have you found a solution?
Has someone found a solution?
> > 4bit 模型是用cuda的算子来进行量化的,没有测试过8bit,可能是会出问题的 > > yes 4 bit is working but 8 bit giving error. > > I have tested cogvlm-chat-hf and it works in 8 bit same code...
@Han230104 So how much VRAM is actually being used in your A6000?
So i have to create a new .py file which consists of my class or i should write another class within that file?