Karl Higley

Results 135 comments of Karl Higley

Yeah, I think that makes sense as the next incremental step from where we are now

Distributed Data Parallel training is something we want to do, but I don't think it's part of this effort to fix the immediate blockers. Does that match what y'all understand,...

Either DistributedDataParallel training is part of the scope of the quick fixes or it isn't, and it sounds like it isn't so we should track that work somewhere (but not...

So, based on our conversation the other day, I'm not sure we actually do need NVT in the ensemble graph in order to send raw data in the request, since...

(If neither of those approaches work, then we should talk about getting this epic issue prioritized on the roadmap. @EvenOldridge would be a good person to talk to about that.)

Updated the title to include the scope of code changes required by the RecSys Demo

The outstanding issue here is that we need the combined training/inference containers to make it possible to run the multi-stage recommender example via one-click deploy.

Duplicate of NVIDIA-Merlin/models#450

Closing https://github.com/NVIDIA-Merlin/models/issues/450 instead