anshul-si

Results 5 issues of anshul-si

**Summary:** During this experiment to integrate the new replicate function into torchtitan, I used https://github.com/pytorch/pytorch/pull/162021, which has not been landed. However, since this is more about making replicate more efficient...

CLA Signed

**Summary:** Currently, whenever we subtract two partial dtensors, we redistribute since linearity is -1 for aten.sub.tensor. However, this is an unnecessary redistribution that can be avoided in similar ways to...

ciflow/trunk
topic: not user facing
ciflow/inductor
module: dtensor

**Summary:** When we add a scalar to Partial dtensor, we don't redistribute causing us to add the scalar to each part of the partial dtensor as reported in https://github.com/pytorch/pytorch/issues/149768, https://github.com/pytorch/pytorch/issues/163193....

ciflow/trunk
topic: not user facing
ciflow/inductor
module: dtensor

**Summary:** While linearities are correct for regular Partial tensors, they don't apply to NormPartial tensors. We fix this by redistributing for all ops necessary when the placement is NormPartial. The...

ciflow/trunk
topic: not user facing
ciflow/inductor
module: dtensor

**Summary:** When users multiply a partial dtensor by a scalar, we should keep the output as a partial dtensor. However, we incorrectly force the partial to be redistributed to replicate...

ciflow/trunk
topic: not user facing
ciflow/inductor
module: dtensor