Andrei Panferov

Results 41 comments of Andrei Panferov

With `distributed=True` you pass the device associated with current worker. For example, if you have 4 _GPUs_, you launch with `--nproc_per_node=4`, extract local worker rank with `LOCAL_RANK = int(os.environ['LOCAL_RANK'])` and...