PyTorch-LIT
PyTorch-LIT copied to clipboard
Lite Inference Toolkit (LIT) for PyTorch
Results
2
PyTorch-LIT issues
Sort by
recently updated
recently updated
newest added
The output of gpt-j is very slow, for a 200 output token generation it takes about 20 minutes, for 2048 it takes more than an hour, this significantly limits any...
Hi, thanks for your effort for making it easy to load and do inference from large models. I tried your code on a gpt-j model with different model file format,...