DeepSpeed-MII
DeepSpeed-MII copied to clipboard
Is there a way for mii to not occupy all the available gpu memory
When I run examples in DeepSpeedExamples to start a server, it occupy all the GPU memory at the beginning. Is it possible to config the max gpu memory that it should occupy? Also what is the recommended way to host two models in the same container?