LongLM
LongLM copied to clipboard
[ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning
Thanks for your contribution on accommondating qwen on self-extend. Qwen1.5 has already been 32k context length. I'm wondering if i can use self-extend to make it to about 100K? Have...
I followed your direction like the below to apply selfextend to llama3 """ [04/19/2024]:💡 We added the support for LLama-3 with transformers==4.40. To use it with transformers==4.40, you may change...
Hello! Thank you for your great work, its amazing how much hard work you put for this algorithm. I just had one question is it possible to integrate this with...
Hi! I have a question that may seem simple, but I think I'm overlooking something. Assume Phi-2's context window is 2K. When we apply a group size ($G_s$) of 4...
In this PR I have added Gemma's example. I have also reported the results in the README.
effeicency -> efficiency