PyTorch-LIT icon indicating copy to clipboard operation
PyTorch-LIT copied to clipboard

Lite Inference Toolkit (LIT) for PyTorch

Results 2 PyTorch-LIT issues
Sort by recently updated
recently updated
newest added

The output of gpt-j is very slow, for a 200 output token generation it takes about 20 minutes, for 2048 it takes more than an hour, this significantly limits any...

Hi, thanks for your effort for making it easy to load and do inference from large models. I tried your code on a gpt-j model with different model file format,...