fwdgrad icon indicating copy to clipboard operation
fwdgrad copied to clipboard

No speed-up in my implementation too

Open LSC527 opened this issue 2 years ago • 3 comments

I implemented this papaer with torch.autograd.forward_ad. However, fwd gradient showed no speed-up compared to fwd+bwd.

LSC527 avatar May 19 '22 02:05 LSC527

It would be interesting for us to see your implementation as well. If you want, you can make a PR to our repo with your code. So we can have multiple implementations available.

DavideTr8 avatar May 23 '22 16:05 DavideTr8

I ran the code from the repository, but I couldn't replicate the results mentioned in the paper, especially regarding the CNN. I used the hyperparameter settings specified in the paper.

May I inquire if there are alternative parameter settings available?

image

LittleWork123 avatar Jan 18 '24 14:01 LittleWork123

Hi, unfortunately we weren't able to reproduce the same results too. The hyperparameters we used are the same reported in the paper, but we don't now if alternative hyperparameters settings are available.

We believe that the difference between our implementation and the official one are due to the fact that they did not use functorch

DavideTr8 avatar Jan 20 '24 16:01 DavideTr8