jiahai-feng

Results 2 comments of jiahai-feng

Thanks a lot for getting back so quickly! I'm wondering if you could kindly point me to a reference explaining how the kernel function in the RNN snippet works. My...

I've encountered the same issue. Instead of changing things at various places in the attention block, I added these four lines to `HookedTransformer.forward` (roughly L546) ``` if attention_mask is not...