singa
singa copied to clipboard
can sparse all-reduce keep efficiency with large number of gpu workers?
in my opinion, when the gpu cluster scaled up to several hundred workers, high sparsification ratios still generate significant communication overheads, which even worst than DenseAllReduce.