transformers
transformers copied to clipboard
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
# What does this PR do? This PR extends VisualQuestionAnsweringPipeline to accept `words` and `boxes` as input, passes them into the tokenizer/model (along with the question), and post-processes their `QuestionAnsweringModelOutput`...
### Feature request Currently, non-Flax models allow `inputs_embeds` instead of `input_ids` (e.g., GPT2) ```python def forward( self, input_ids: Optional[torch.LongTensor] = None, ... inputs_embeds: Optional[torch.FloatTensor] = None, ... ) -> Union[Tuple,...
# What does this PR do? There are two issues being attacked: 1. Always creating a fresh pool within `Wav2Vec2ProcessorWithLM.batch_decode` generates a big overhead if it's called multiple times (this...
# What does this PR do? Currently when we want to register a new config+tokenizer+model, per [the instructions](https://huggingface.co/docs/transformers/model_doc/auto), it seems we should do the following: ``` from transformers import AutoConfig,...
# What does this PR do? __Context__ It was reported in optimum https://github.com/huggingface/optimum/issues/305 that the training on DeBERTa with optimum.onnxruntime.ORTTrainer is broken. After investigation, the break comes from two causes:...
# What does this PR do? 'True' losses should be computed in Flax examples, as [discussed](https://github.com/huggingface/transformers/pull/18297#discussion_r931971230) with @sanchit-gandhi. ## Who can review? cc @sanchit-gandhi @patrickvonplaten
Update parameter counts of BLOOM models. The original counts were incorrect & have already been updated on the hub. I can't add reviewers, but @younesbelkada @thomasw21 may want to review...
# Adding TensorFlow version of GroupViT This PR adds the TensorFlow version of [GroupViT](https://github.com/NVlabs/GroupViT). ## Before submitting - [ ] This PR fixes a typo or improves the docs (you...
# What does this PR do? This is a feature addition to LayoutLMV3, a follow up to the same feature added to [LayoutLMV2's feature extractor ](https://github.com/huggingface/transformers/pull/17733) Giving user option to...
Hey @NielsRogge I found an inconsistency in the documentation and the code for configuration of GroupViT. The default for `num_output_groups` is `[64, 8, 8]` (notice the last element in the...