Uri Alon
Uri Alon
Thanks @jordancole21 , We are currently working on supporting decoder models such as LLaMA, so I am keeping this issue open. If you manage to solve this in the meantime,...
Hi @TrieuLe0801 , Running on GPU or CPU is done in the exact same way! Did you try?
Hi @TrieuLe0801 , There is no specific support for M1 - because it should be supported by the `torch` library. Everything is supposed to work the same way. Please let...
Hi @Muxv , Thank you for your interest in our work! It is currently not fully supported. I tried to run with multiple GPUs once, but I got errors that...
Hi Hossam, I don't think I implemented this function myself, I think I copied it from Huggingface's language modeling example. If I remember correctly, it's just more efficient than padding,...
Hi Hossam, Thank you for your interest in our work. I believe that you need to rebuild the KNN datastore specifically for distill-GPT. Have you done that? Best, Uri On...
I just replied to you in a different thread, let me if anything is still unclear. On Sun, Oct 20, 2024 at 13:35 Hossam Amer ***@***.***> wrote: > Hi Uri,...
Hi Rubin, Thank you for your interest in our work. Does it still happen when you use our datastore and our index? Best, Uri On Wed, Dec 4, 2024 at...
Hi folks, Thank you for your interest in our work. Unfortunately, this codebase is 4 years old. I don't have the capacity to investigate why KNN-GPU gives different results than...
Hi, Thank you for your interest in our work! Did you try running with our datastore, our index etc? (Without tuning anything) On Wed, Dec 4, 2024 at 11:24 He...