jiahai-feng
Results
2
comments of
jiahai-feng
Thanks a lot for getting back so quickly! I'm wondering if you could kindly point me to a reference explaining how the kernel function in the RNN snippet works. My...
I've encountered the same issue. Instead of changing things at various places in the attention block, I added these four lines to `HookedTransformer.forward` (roughly L546) ``` if attention_mask is not...