wanda icon indicating copy to clipboard operation
wanda copied to clipboard

Some question about the code

Open liuxiaozhu01 opened this issue 11 months ago • 2 comments

Hi! Thanks for your great work!

Im a little confused about the implementation. Your simple and efficient work only requires once forward caculate to get the activation of each layer. This line seems means that the forward caculate is excuted in sparse network, which also means the input of next layer is caculated by the current sparse layer because current layer is already masked.

Im wondering whether the "masked forward" is necessary, and I notice that if this operation is canceled, the result can be better in some conditions.

liuxiaozhu01 avatar Mar 11 '24 10:03 liuxiaozhu01