iefgnoix
iefgnoix
I created a [PR](https://github.com/pytorch/xla/pull/3823) but there is lingering issue: CircleCI tests fails (in particular, test_index_copy_xla_* and test_put_xla_*).
Question question: I notice in the class Flip there is a private member `::std::vector dims;`. For me, most of the argument I passed to a XlaNode constructor are converted to...
> > Quick question: I notice in the class Flip there is a private member `::std::vector dims;`. For me, most of the argument I passed to a XlaNode constructor are...
In response to the first comment, at https://github.com/pytorch/pytorch/blob/d39e9c1e9087069fa774b0e3eb47e04750edca88/c10/core/SymIntNodeImpl.h#L85, I changed to a more specific error string, such as "str() NYI". Then I rebuilt pytorch and run the commands: ``` >>>...
Okay, I got the c++ stacktrace: ``` (pytorch) root@t1v-n-cf794107-w-0:/# python3 Python 3.8.8 (default, Apr 13 2021, 19:58:26) [GCC 7.3.0] :: Anaconda, Inc. on linux Type "help", "copyright", "credits" or "license"...
Thanks for the response. In that case, can we add a GPU XLA tests through post-submit? That will help catch issues earlier.
> Certainly. > > IIUC all that's needed to get this going is to install `torch_xla` currently with the state of accelerate, correct? If so then we just need to...
Yes. You can use this: PJRT_DEVICE=CUDA python ``` import torch, torch_xla import torch_xla.core.xla_model as xm t1 = torch.randn(1, 128, device='cpu') t2 = torch.randn(1, 128, device='cpu') xt1 = t1.to(xm.xla_device()) xt2 =...
> BTW just noticing this, we should eventually change the logic so PJRT_DEVICE is auto-set if multi-gpu is enabled through the config file + torch_xla is available. Completed agreed. >...