Yan Wang

Results 27 issues of Yan Wang

Add a prim operator to support nvfuser `fd.add_output(output, alias_input)`, which allows in-place updating of the input tensor. Related context is we want to use it to update the running stats...

Support batch norm: - Add backward support - Add nvFuser support - Type promotion of input/weight/bias are handled in thunder - Running stats are left as-is, and let nvfuserex/torchex to...

Before submitting - [ ] Was this discussed/approved via a Github issue? (no need for typos and docs improvements) - [ ] Did you read the [contributor guideline](https://github.com/Lightning-AI/pytorch-lightning/blob/main/.github/CONTRIBUTING.md), Pull Request...

distributed

Background: Currently, the `mask` calculated for `dropout` in the forward trace is passed to the backward trace, which takes too much memory. We want Thunder to save the seed and...

design required

## 🚀 Feature Add the benchmark for ResNet50 cc @crcrpar

enhancement
benchmarking

Before submitting - [ ] Was this discussed/approved via a Github issue? (no need for typos and docs improvements) - [ ] Did you read the [contributor guideline](https://github.com/Lightning-AI/pytorch-lightning/blob/main/.github/CONTRIBUTING.md), Pull Request...

## 🐛 Bug When I add an operator that returns numbers, the values in the saved_for_backward are the compile time value defined in `Symbol.meta`, not the real value computed at...

bug
help wanted
transforms
dynamic constraints

## 🐛 Bug The same function outputs different values when the input tensor is the same but `requires_grad` is `True/False`. note: if change the last line in `func` to be...

bug
operators
randomness

Before submitting - [ ] Was this discussed/approved via a Github issue? (no need for typos and docs improvements) - [ ] Did you read the [contributor guideline](https://github.com/Lightning-AI/pytorch-lightning/blob/main/.github/CONTRIBUTING.md), Pull Request...

## 🐛 Bug ```py import torch import torch.nn.functional as F import thunder a = torch.randn(1, 64, 112, 112).cuda().requires_grad_() def func(a): return F.max_pool2d(a, 3, 2, 1, 1, False, False) # t79:...

bug
operators