Matt

Results 203 comments of Matt

@sgugger Absolutely agree on the potential for small bugs, but in the case of model porting doesn't the CI test equivalence with the PT original? If the model accepts various...

I'm still working on the prompt!!!!!

Interesting! Can you document exactly what error you got with the compile step and what code you ran to cause them?

Yes please! Ideally if you could give us some minimal code that reproduces the issue, that would make it much easier for us to track it down. Also, sorry for...

Ah, I see! The issue here is caused by some specific behaviour of the SegFormer models when using inputs of this resolution. The model outputs are actually at a lower...

Yeah, refactoring our TF models to make `summary()` more usable is absolutely on the list! Unfortunately it's quite a big list, but it's definitely there.

The TF port is mostly complete now and tests are passing locally - I just need to go around updating docs and auto classes and so on. The main code...

Got through a lot of the comments today, but I have a couple of other things to do - will try to finish them tomorrow!

The last remaining big issue is that some of the pt-tf equivalence tests fail when weights don't match up between models. This is caused by the cross-attention weights not being...

The issue seems to be that in all of our other models, cross-attention layers are only added when `config.add_cross_attention` is True, but in the case of BLIP it only checks...