InfLLM icon indicating copy to clipboard operation
InfLLM copied to clipboard

`Position Emb` and `Chunk size`

Open liyucheng09 opened this issue 9 months ago • 0 comments

Great job, I found two problems when trying to reproduce the paper's results.

  1. The same positiona embedding was used for all context memory units as explained in the paper. But I found in code implementation, there seems no use of position embedding for cached Ks at all?

  2. Why chunk size? The proposed method does the attention block by block, which (I think) wouldn't cause OOM errors even without the chunking trick in decoding. But I found it fail to process 100K text without setting chunk size, while using flash attn is totaly fine in such circumstances.

liyucheng09 avatar May 05 '24 05:05 liyucheng09