Bryce Meyer
Bryce Meyer
Someone opened a PR last night for LLaMA 3.1, but I am sure we can find one for you to add. Is there a specific model you are interested in?...
@deven367 That sounds perfect to me! Ping me on Slack, and we can discuss further if you like. The HookedEncoderDecord is a relatively newer addition to TransformerLens, and the only...
Sorry for the delay on getting this in, and thanks for bumping me. I updated it to the most recent version of TransformerLens, and that seems to have caused some...
That is very odd. The token in there should be mine, and I should have access to all of those models. Can you remove your change that allows it to...
TransformerLens does not currently support Baichuan. Looking at their page on HuggingFace, it appears to be pretty similar to LLaMA, which means that it should be relatively easy to add,...
Thank you very much for doing the work here to get it to work in TransformerLens! Your implementation from a glance seems correct. However, it's hard to say if the...
TransformerLens just started officially supporting vision models in the last month. There is a way to basically make it happen without needing to wait for a specific model to be...
Thank you for being so thorough on this. We are in the middle of working out some tools that will make benchmarking models easier, and I think that is going...
@jkminder If you are looking into this, the thing to do is check the implementation of the model in `transformers` (https://github.com/huggingface/transformers/tree/main/src/transformers/models), and check how the MLPs work there. I found...
@jkminder Sorry for the delay on getting back to you. we have some pretty major changes coming to existing components in the same vein as what you are bringing up....