clarinevong
clarinevong
Thanks for the reply and the different suggestions. I was actually able to recreate the issue on Linux with - Optimum Version: 1.18.0 - Python Version: 3.10.12 - Torch Version:...
Yes of course ``` aiohttp==3.9.3 aiosignal==1.3.1 async-timeout==4.0.3 attrs==23.2.0 certifi==2024.2.2 charset-normalizer==3.3.2 coloredlogs==15.0.1 datasets==2.18.0 dill==0.3.8 filelock==3.13.4 flatbuffers==24.3.25 frozenlist==1.4.1 fsspec==2024.2.0 huggingface-hub==0.22.2 humanfriendly==10.0 idna==3.6 Jinja2==3.1.3 joblib==1.4.0 MarkupSafe==2.1.5 mpmath==1.3.0 multidict==6.0.5 multiprocess==0.70.16 networkx==3.3 numpy==1.26.4 nvidia-cublas-cu12==12.1.3.1 nvidia-cuda-cupti-cu12==12.1.105...
Hello, is there any news on this issue? I encounter a similar issue. I have installed both `LLamaSharp.Backend.CPU` and `LLamaSharp.Backend.Cuda12.Windows` (0.18.0 versions). Following th README I added the following line...
> > When I load a model on the CPU with GpuLayerCount equals to 0, the cuda backend is loaded > > That's how it's meant to work - if...