MADE-with-PyTorch
MADE-with-PyTorch copied to clipboard
Companion layers not resampled
Thank you for this implementation of MADE. But I don't understand why companion layers are needed in the model.
Section 4.3 of the paper says:
instead of sampling the values of ml(k) for all units and layers once and for all before training, we actually resample them for each training example or minibatch.
In your implementation it seems ml(k) is chosen only once before training. If that's the case, how is it different from other linear layers?
Thanks in advance.