MLServer
MLServer copied to clipboard
build(deps-dev): bump transformers from 4.41.2 to 4.42.4 in /runtimes/huggingface
Bumps transformers from 4.41.2 to 4.42.4.
Release notes
Sourced from transformers's releases.
Patch release v4.42.4
Mostly gemma2 support FA2 softcapping!
but also fix the sliding window for long context and other typos.
- [Gemma2] Support FA2 softcapping (#31887) by
@ArthurZucker- [ConvertSlow] make sure the order is preserved for addedtokens (#31902) by
@ArthurZucker- Fixes to alternating SWA layers in Gemma2 (#31775) by
@turboderp- Requires for torch.tensor before casting (#31755) by
@echarlaixWas off last week could not get this out, thanks all for your patience 🥳
Patch release v4.42.3
Make sure we have attention softcapping for "eager" GEMMA2 model
After experimenting, we noticed that for the 27b model mostly, softcapping is a must. So adding it back (it should have been there, but an error on my side made it disappear) sorry all! 😭
- Gemma capping is a must for big models (#31698)
Patch release v4.42.2
Patch release
Thanks to our 2 contributors for their prompt fixing mostly applies for training and FA2!
- Fix Gemma2 4d attention mask (#31674) by
@hiyouga- don't zero out the attention_mask when using sliding window with flash attention (#31670) by
@winglianv4.42.1: Patch release
Patch release for commit:
- [HybridCache] Fix get_seq_length method (#31661)
v4.42.0: Gemma 2, RTDETR, InstructBLIP, LLAVa Next, New Model Adder
New model additions
Gemma-2
The Gemma2 model was proposed in Gemma2: Open Models Based on Gemini Technology and Research by Gemma2 Team, Google. Gemma2 models are trained on 6T tokens, and released with 2 versions, 2b and 7b.
The abstract from the paper is the following:
This work introduces Gemma2, a new family of open language models demonstrating strong performance across academic benchmarks for language understanding, reasoning, and safety. We release two sizes of models (2 billion and 7 billion parameters), and provide both pretrained and fine-tuned checkpoints. Gemma2 outperforms similarly sized open models on 11 out of 18 text-based tasks, and we present comprehensive evaluations of safety and responsibility aspects of the models, alongside a detailed description of our model development. We believe the responsible release of LLMs is critical for improving the safety of frontier models, and for enabling the next wave of LLM innovations
- Add gemma 2 by
@ArthurZuckerin #31659RTDETR
... (truncated)
Commits
fc35907v4.42.4e002fcd[Gemma2] Support FA2 softcapping (#31887)2e43416[ConvertSlow] make sure the order is preserved for addedtokens (#31902)c43fd9dFixes to alternating SWA layers in Gemma2 (#31775)0be998bRequires for torch.tensor before casting (#31755)b7ee1e8v4.42.3da50b41Gemma capping is a must for big models (#31698)086c74ev4.42.28691867Fix Gemma2 4d attention mask (#31674)7edc993don't zero out the attention_mask when using sliding window with flash attent...- Additional commits viewable in compare view
Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.
Dependabot commands and options
You can trigger Dependabot actions by commenting on this PR:
@dependabot rebasewill rebase this PR@dependabot recreatewill recreate this PR, overwriting any edits that have been made to it@dependabot mergewill merge this PR after your CI passes on it@dependabot squash and mergewill squash and merge this PR after your CI passes on it@dependabot cancel mergewill cancel a previously requested merge and block automerging@dependabot reopenwill reopen this PR if it is closed@dependabot closewill close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually@dependabot show <dependency name> ignore conditionswill show all of the ignore conditions of the specified dependency@dependabot ignore this major versionwill close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)@dependabot ignore this minor versionwill close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)@dependabot ignore this dependencywill close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)