kaito
kaito copied to clipboard
chore: bump accelerate from 0.23.0 to 0.30.1 in /presets/inference/text-generation
Bumps accelerate from 0.23.0 to 0.30.1.
Release notes
Sourced from accelerate's releases.
v0.30.1: Bugfixes
Patchfix
- Fix duplicate environment variable check in multi-cpu condition thanks to
@yhna940in huggingface/accelerate#2752- Fix issue with missing values in the SageMaker config leading to not being able to launch in huggingface/accelerate#2753
- Fix CPU OMP num threads setting thanks to
@jiqing-fengin huggingface/accelerate#2755- Fix FSDP checkpoint unable to resume when using offloading and sharded weights due to CUDA OOM when loading the optimizer and model huggingface/accelerate#2762
- Fixed the problem of incorrect conditional judgment statement when configuring enable_cpu_affinity thanks to
@statelesshzin huggingface/accelerate#2748- Fix stacklevel in logging to log the actual user call site (instead of the call site inside the logger wrapper) of log functions thanks to
@luowyangin huggingface/accelerate#2730- Fix support for multiple optimizers when using LOMO thanks to
@younesbelkadain huggingface/accelerate#2745Full Changelog: https://github.com/huggingface/accelerate/compare/v0.30.0...v0.30.1
v0.30.0: Advanced optimizer support, MoE DeepSpeed support, add upcasting for FSDP, and more
Core
- We've simplified the
tqdmwrapper to make it fully passthrough, no need to havetqdm(main_process_only, *args), it is now justtqdm(*args)and you can pass inis_main_processas a kwarg.- We've added support for advanced optimizer usage:
- Schedule free optimizer introduced by Meta by
@muellerzrin huggingface/accelerate#2631- LOMO optimizer introduced by OpenLMLab by
@younesbelkadain huggingface/accelerate#2695- Enable BF16 autocast to everything during FP8 and enable FSDP by
@muellerzrin huggingface/accelerate#2655- Support dataloader send_to_device calls to use non-blocking by
@drheadin huggingface/accelerate#2685- allow gather_for_metrics to be more flexible by
@SunMarcin huggingface/accelerate#2710- Add
cannversion info to command accelerate env for NPU by@statelesshzin huggingface/accelerate#2689- Add MLU rng state setter by
@ArthurinRUCin huggingface/accelerate#2664- device agnostic testing for hooks&utils&big_modeling by
@statelesshzin huggingface/accelerate#2602Documentation
- Through collaboration between
@fabianlim(lead contribuitor),@stas00,@pacman100, and@muellerzrwe have a new concept guide out for FSDP and DeepSpeed explicitly detailing how each interop and explaining fully and clearly how each of those work. This was a momumental effort by@fabianlimto ensure that everything can be as accurate as possible to users. I highly recommend visiting this new documentation, available here- New distributed inference examples have been added thanks to
@SunMarcin huggingface/accelerate#2672- Fixed some docs for using internal trackers by
@brentyiin huggingface/accelerate#2650DeepSpeed
- Accelerate can now handle MoE models when using deepspeed, thanks to
@pacman100in huggingface/accelerate#2662- Allow "auto" for gradient clipping in YAML by
@regisssin huggingface/accelerate#2649- Introduce a
deepspeed-specific Docker image by@muellerzrin huggingface/accelerate#2707. To use, pull thegpu-deepspeedtagdocker pull huggingface/accelerate:cuda-deepspeed-nightlyMegatron
- Megatron plugin can support NPU by
@zhangsheng377in huggingface/accelerate#2667Big Modeling
- Add strict arg to load_checkpoint_and_dispatch by
@SunMarcin huggingface/accelerate#2641Bug Fixes
- Fix up state with xla + performance regression by
@muellerzrin huggingface/accelerate#2634- Parenthesis on xpu_available by
@muellerzrin huggingface/accelerate#2639- Fix
is_train_batch_mintype in DeepSpeedPlugin by@yhna940in huggingface/accelerate#2646- Fix backend check by
@jiqing-fengin huggingface/accelerate#2652- Fix the rng states of sampler's generator to be synchronized for correct sharding of dataset across GPUs by
@pacman100in huggingface/accelerate#2694- Block AMP for MPS device by
@SunMarcin huggingface/accelerate#2699- Fixed issue when doing multi-gpu training with bnb when the first gpu is not used by
@SunMarcin huggingface/accelerate#2714
... (truncated)
Commits
b52803dRelease: v0.30.1d8c8634LOMO / FIX: Support multiple optimizers (#2745)d58a05aFix stacklevel inloggingto log the actual user call site (instead of the ...4aa2eafFixed the problem of incorrect conditional judgment statement when configurin...4a92761Enable sharded cpu resume (#2762)0f0a856fix cpu omp num threads set (#2755)0292728Fix sagemaker config (#2753)e0ef0c5Fix duplicate environment variable check in multi-cpu condition (#2752)989cc50Release: v0.30.0060361fFix tests on main (#2739)- Additional commits viewable in compare view
Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR:
@dependabot rebasewill rebase this PR@dependabot recreatewill recreate this PR, overwriting any edits that have been made to it@dependabot mergewill merge this PR after your CI passes on it@dependabot squash and mergewill squash and merge this PR after your CI passes on it@dependabot cancel mergewill cancel a previously requested merge and block automerging@dependabot reopenwill reopen this PR if it is closed@dependabot closewill close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually@dependabot show <dependency name> ignore conditionswill show all of the ignore conditions of the specified dependency@dependabot ignore this major versionwill close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)@dependabot ignore this minor versionwill close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)@dependabot ignore this dependencywill close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)