transformers
                                
                                 transformers copied to clipboard
                                
                                    transformers copied to clipboard
                            
                            
                            
                        đ¤ Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
### Feature request Add quantization_config feature to AutoModelForCausalLM from config . I am trying to pretrain a model from scratch and use bits and bytes so that It can be...
# What does this PR do? This PR does the following: 1. Add a new field to `TrainerArguments.AcceleratorConfig`, `use_stateful_dataloader` which when set to true passes through `use_stateful_dataloader` to the DataLoaderConfiguration...
### System Info - `transformers` version: 4.45.2 - Platform: Linux-5.17.15-051715-generic-x86_64-with-glibc2.35 - Python version: 3.10.12 - Huggingface_hub version: 0.24.3 - Safetensors version: 0.4.3 - Accelerate version: 0.33.0 - Accelerate config: not...
# What does this PR do?
### System Info huggingface-hub-0.25.2 tokenizers-0.20.1 transformers-4.45.2 ### Who can help? @ArthurZucker @itazap ### Information - [ ] The official example scripts - [X] My own modified scripts ### Tasks -...
Current status for AutoTokenizer with fast=True: 1. checks tokenizer_config.json if tokenizer_class name ends with Fast 2. if not, load a slow tokenizer (This PR):â¨(unchanged) 1. checks tokenizer_config.json if tokenizer_class ...
### Feature request I want to add the ability to use GGUF BERT models in transformers. Currently the library does not support this architecture. When I try to load it,...
One unusual change this time: The TokenClassificationPipeline put the docstrings in an odd place, with a decorator. As far as I can tell, this isn't necessary at all - I...
The chat template docs describe using generation prompts, but I realize it's a little unclear for model authors after reading #34209. This PR adds a section on writing generation prompts.
# What does this PR do? It passes the model and the optimizer to accelerate.prepare in order to enable fp8 mixed precision, if any. Fixes #34024 ## Who can review?...