ReCU
ReCU copied to clipboard
Pytorch implementation of our paper accepted by ICCV 2021 -- ReCU: Reviving the Dead Weights in Binary Neural Networks http://arxiv.org/abs/2103.12369
i am wondering in your paper why use latent full precision weights to calculate information entropy rather than binarized weights? It seems make no sense considering latent weights.
During evaluation i get following errors Traceback (most recent call last): File "/home/ahmsoy00/Projects/DATE23_Invited_Variational_Bayes/classification/CIFAR-10/ReCU/main.py", line 319, in main() File "/home/ahmsoy00/Projects/DATE23_Invited_Variational_Bayes/classification/CIFAR-10/ReCU/main.py", line 118, in main val_loss, val_prec1, val_prec5 = validate(val_loader, model, criterion,...
Appreciate your excellent work! I look up the code in **binarized_modules.py** and find that **torch.clamp()** is applied to constrain weights within ±Q_tau. In backwards, this approach stops "dead weights"(which >=Q_tau...
For ResNet in cifar experiment , the shortcut/downsampling is binarized to [-1,+1] via BinarizeConv2d. However in ImageNet experiment , the shortcut/downsampling remains the fullprecision approach via nn.Conv2d. Is this the...
Hello developers, Thank you for sharing your code. Would you mind sharing also the full instructions to reproduce your results with the training set from ReActNet? In particular, did you...