Phil
Phil
Thanks for the detailed investigation and for finding a workaround, really appreciate it! We realize that at the moment the model is quite memory hungry and activation tensor allocations depend...
I've seen segfaults before as well, so while not common it's also not uncommon :) Back then all I could find was "that can have many reasons", like the ones...
Closing due to inactivity.
Note: https://github.com/PriorLabs/TabPFN-private/pull/227 Also needs to be on the server before we can switch to the gated version (otherwise the checkpoint cleaned for deployment fails on load)
Ah sorry, the linked PR is already the copy over from the public repo, so cutting over should be good!
Thanks for reporting! Would you mind trying the workaround suggested in https://github.com/PriorLabs/TabPFN/issues/404 of calling `torch.cuda.empty_cache()` before each predict call and reporting back if this fixes things?
Ah, sorry, I missed that :) Could you share which dtype you were using?