Karl Higley
Karl Higley
Yeah, I think that makes sense as the next incremental step from where we are now
Distributed Data Parallel training is something we want to do, but I don't think it's part of this effort to fix the immediate blockers. Does that match what y'all understand,...
Either DistributedDataParallel training is part of the scope of the quick fixes or it isn't, and it sounds like it isn't so we should track that work somewhere (but not...
I don't think we should split the issue, let's just target this for 22.09
So, based on our conversation the other day, I'm not sure we actually do need NVT in the ensemble graph in order to send raw data in the request, since...
(If neither of those approaches work, then we should talk about getting this epic issue prioritized on the roadmap. @EvenOldridge would be a good person to talk to about that.)
Updated the title to include the scope of code changes required by the RecSys Demo
The outstanding issue here is that we need the combined training/inference containers to make it possible to run the multi-stage recommender example via one-click deploy.
Duplicate of NVIDIA-Merlin/models#450
Closing https://github.com/NVIDIA-Merlin/models/issues/450 instead