MLServer
MLServer copied to clipboard
GPU support for custom inference runtimes in MLServer
I am trying to use GPU in my custom inference endpoint built using MLserver. I am unable to load the model on gpu. Can you please let me know if this is possible or not?
Thank you