gpetters94

Results 60 comments of gpetters94

> I have recently pushed some fixes in this diff: #83092 > > You will need to rebase on top of it. The easiest way to do so would be...

@silvasean I'm picking this one up - how would that logic generalize to TupleConstruct chains? For example, in OPT the final return is of a prim.TupleConstruct, but the arguments of...

If we're just looking for the tensor output then `self.forward()[0]` should work. I'm working on a refactor of the model to see if we can get that to pass.

It's a tensor that is the "last hidden state" as per [this](https://huggingface.co/docs/transformers/model_doc/opt#transformers.OPTModel.forward). I'm not familiar with how NLP models work, so I don't know if the other returns are something...

> @gpetters94 Could you please resolve the conflicts? Done.

It's holding up U-Net support, but it isn't the only thing blocking that.

@ZihengJiang @silvasean I made a small change in the shape logic (I forgot that two of the dims needed transposing) so I need to make another [PR](https://github.com/pytorch/pytorch/pull/83557) in Pytorch that...

@vivekkhandelwal1 @silvasean The upstream shape code is merged, so this one should be good to go.