Yassine Alouini
Yassine Alouini
@gurkirt since I haven't seen your data or used it, I guess the best way to know is by experimenting. Regarding the alignment, it is using this [code](https://gist.github.com/ranftlr/45f4c7ddeb1bbb88d606bc600cab6c8d) (check the...
Thanks for this great guide @jdsgomes. There is a small typo here: [ConvNormActication](https://github.com/pytorch/vision/blob/47bd962069ba03f753e7ba711cb825317be0b00a/torchvision/ops/misc.py#L104) (should be Activation)
Perfect, I will work on this today but mostly next week and the week after. Will let you know how my progress goes. 👌
I have started the implementation. Seems like a big chunk but excited to do it. :ok_hand: I have found this huggingface implementation, could be useful as another inspiration: https://huggingface.co/docs/transformers/main/model_doc/mobilevit. [EDIT]...
Hello @datumbox, thanks for checking. So far, so good. It is taking a bit longer since I only had one day of working on it and it is paused for...
Some update @datumbox: I will have some free time for the upcoming few days and should make some progress. Will let you know how it goes. 👌
By the way, what is the PyTorch and TorchVision policies for the usage of einops? 🤔
I don't think it is irreplaceable, just wanted to check what is the best practice in torchvision. 👌 I will code everything using PyTorch and existing TorchVision code.
One additional question regarding the `TransformerEncoder`: should I reimplement it or should I re-use the one from `vision_transformer.py` (i.e. `EncoderBlock`)? I was planning to copy-paste the code first, adapt it...
Is there a new URL for the pretrained model? Or maybe it has been removed from some other reasons? Any help is welcome. :)