atlas icon indicating copy to clipboard operation
atlas copied to clipboard

Save and load compressed index

Open xudongguan202 opened this issue 1 year ago • 2 comments

Hi,

I got a total 755G index saved in my disk after encoding the whole wiki passage. The large index takes huge storage and long time to load to GPU. However, it requires less than 100G after loading to GPU, which could be the index compression mentioned in your paper. Is it possible to save and load the compressed index for better time and storage consumption?

xudongguan202 avatar Sep 05 '23 03:09 xudongguan202

Yeh, I met the same situation like u I loaded the embeddings and passages of precomputing, but when I run evaluate.py, it always shows cuda out of memory while loading. I have an A100 80G GPU memory. How much GPU memory is needed at least to load the precomputing embedding? And I also encountered the problem of extremely slowly loading. Is there any way to optimize it?

Vincent-ch99 avatar Sep 07 '23 11:09 Vincent-ch99

@Vincent-ch99 At least 2x A100 80G GPUs are required to run the evaluation using the default configurations from my experiment.

The extremely slow loading could be due to index compression of saved index, as mentioned in the paper. Therefore, I am currently trying to find a way to save and load the compressed index for faster loading and light storage.

xudongguan202 avatar Sep 07 '23 11:09 xudongguan202