How to load the model correctly without running any prediction ? ( Just warming )
I suppose calling the /predictions with falsy value can do the trick but it feels dirty. Is there a nicer way to pre warm a model ? The idea would be to pre warm when the user select such model but before launching the generation, potentially never launching it indeed, but i'm okay with this.
Thanks
I do believe that cog calls the runners setup once the FastAPI app has started up, so it should be warm, can you let me know anything about your setup that might be causing it not to honour this contract?
Well the way i trigger the docker mounting is by calling the full url with /prediction, maybe calling it without would work, but i was asking for the 'clean way'. Maybe its clean this way ? Would just pinging the url on the machine prevents it to cool down too ?
Ah I see, I believe what you may want is something like /docs might work for you here.