transformers
transformers copied to clipboard
π€ Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
# What does this PR do? Fixes #37171 and continuation of #37392
# What does this PR do? Improves `eager_attention_forward` and `sdpa_attention_forward` in the case when `query.shape[1] > key.shape[1]`. This happens in GQA (grouped query attention), and in particular in multi-head latent...
@ArthurZucker, pls help review, thx very much.
# What does this PR do? Fixes https://github.com/huggingface/transformers/issues/37051 Approach is to support uneven sharding and seek segments of data that mimics torch.chunk since torch.chunk is the style of sharding adopted...
Fixes #38468 ### Problem Description This PR addresses the `AssertionError: "Torch not compiled with CUDA enabled"` that occurs when attempting to load models using `device_map="auto"` on systems with Ascend NPU...
# What does this PR do? Draft PR to add Dust3R Cc @raf-fonseca
### System Info - `transformers` version: 4.53.0.dev0 - Platform: Linux-5.10.0-34-cloud-amd64-x86_64-with-glibc2.31 - Python version: 3.9.2 - Huggingface_hub version: 0.32.2 - Safetensors version: 0.4.5 - Accelerate version: 1.7.0 - Accelerate config: not...
Description: This pull request fixes minor typos in comments across two files: - In `modeling_swin.py`, the word "disible" was corrected to "divisible" for clarity in the padding comment. - In...
# This PR adds support for the Quartet QAT method. The goal of this PR is to integrate inference and training support for the [Quartet QAT method](https://arxiv.org/abs/2505.14669). That would allow...
### System Info - `transformers` version: 4.31.0 - Platform: macOS-13.4.1-arm64-arm-64bit - Python version: 3.11.4 - Huggingface_hub version: 0.15.1 - Safetensors version: 0.3.1 - Accelerate version: 0.21.0 - Accelerate config: not...