Skip to content

Commit

Permalink
build(deps): bump transformers from 4.41.2 to 4.42.3 (#883)
Browse files Browse the repository at this point in the history
Bumps [transformers](https://github.com/huggingface/transformers) from
4.41.2 to 4.42.3.
<details>
<summary>Release notes</summary>
<p><em>Sourced from <a
href="https://github.com/huggingface/transformers/releases">transformers's
releases</a>.</em></p>
<blockquote>
<h2>Patch release v4.42.3</h2>
<h1>Make sure we have attention softcapping for &quot;eager&quot; GEMMA2
model</h1>
<p>After experimenting, we noticed that for the 27b model mostly,
softcapping is a must. So adding it back (it should have been there, but
an error on my side made it disappear) sorry all! 😭</p>
<ul>
<li>Gemma capping is a must for big models (<a
href="https://redirect.github.com/huggingface/transformers/issues/31698">#31698</a>)</li>
</ul>
<h2>Patch release v4.42.2</h2>
<h1>Patch release</h1>
<p>Thanks to our 2 contributors for their prompt fixing mostly applies
for training and FA2!</p>
<ul>
<li>Fix Gemma2 4d attention mask (<a
href="https://redirect.github.com/huggingface/transformers/issues/31674">#31674</a>)
by <a href="https://github.com/hiyouga"><code>@​hiyouga</code></a></li>
<li>don't zero out the attention_mask when using sliding window with
flash attention (<a
href="https://redirect.github.com/huggingface/transformers/issues/31670">#31670</a>)
by <a
href="https://github.com/winglian"><code>@​winglian</code></a></li>
</ul>
<h2>v4.42.1: Patch release</h2>
<p>Patch release for commit:</p>
<ul>
<li>[HybridCache] Fix get_seq_length method (<a
href="https://redirect.github.com/huggingface/transformers/issues/31661">#31661</a>)</li>
</ul>
<h2>v4.42.0: Gemma 2, RTDETR, InstructBLIP, LLAVa Next, New Model
Adder</h2>
<h2>New model additions</h2>
<h3>Gemma-2</h3>
<p>The Gemma2 model was proposed in <a
href="https://blog.google/technology/developers/Gemma2-open-models/">Gemma2:
Open Models Based on Gemini Technology and Research</a> by Gemma2 Team,
Google.
Gemma2 models are trained on 6T tokens, and released with 2 versions, 2b
and 7b.</p>
<p>The abstract from the paper is the following:</p>
<p><em>This work introduces Gemma2, a new family of open language models
demonstrating strong performance across academic benchmarks for language
understanding, reasoning, and safety. We release two sizes of models (2
billion and 7 billion parameters), and provide both pretrained and
fine-tuned checkpoints. Gemma2 outperforms similarly sized open models
on 11 out of 18 text-based tasks, and we present comprehensive
evaluations of safety and responsibility aspects of the models,
alongside a detailed description of our model development. We believe
the responsible release of LLMs is critical for improving the safety of
frontier models, and for enabling the next wave of LLM
innovations</em></p>
<p><img
src="https://github.com/huggingface/transformers/assets/30755778/798b25f4-485a-4b60-abe5-af612def209b"
alt="image" /></p>
<ul>
<li>Add gemma 2 by <a
href="https://github.com/ArthurZucker"><code>@​ArthurZucker</code></a>
in <a
href="https://redirect.github.com/huggingface/transformers/issues/31659">#31659</a></li>
</ul>
<h3>RTDETR</h3>
<p>The RT-DETR model was proposed in <a
href="https://arxiv.org/abs/2304.08069">DETRs Beat YOLOs on Real-time
Object Detection</a> by Wenyu Lv, Yian Zhao, Shangliang Xu, Jinman Wei,
Guanzhong Wang, Cheng Cui, Yuning Du, Qingqing Dang, Yi Liu.</p>
<p>RT-DETR is an object detection model that stands for “Real-Time
DEtection Transformer.” This model is designed to perform object
detection tasks with a focus on achieving real-time performance while
maintaining high accuracy. Leveraging the transformer architecture,
which has gained significant popularity in various fields of deep
learning, RT-DETR processes images to identify and locate multiple
objects within them.</p>
<p><img
src="https://github.com/huggingface/transformers/assets/30755778/78b096d4-2686-41cb-9fdd-1cd517722fd3"
alt="image" /></p>
<ul>
<li>New model support RTDETR by <a
href="https://github.com/SangbumChoi"><code>@​SangbumChoi</code></a> in
<a
href="https://redirect.github.com/huggingface/transformers/issues/29077">#29077</a></li>
</ul>
<h3>InstructBlip</h3>
<p>The InstructBLIP model was proposed in <a
href="https://arxiv.org/abs/2305.06500">InstructBLIP: Towards
General-purpose Vision-Language Models with Instruction Tuning</a> by
Wenliang Dai, Junnan Li, Dongxu Li, Anthony Meng Huat Tiong, Junqi Zhao,
Weisheng Wang, Boyang Li, Pascale Fung, Steven Hoi. InstructBLIP
leverages the <a
href="https://huggingface.co/docs/transformers/main/en/model_doc/blip2">BLIP-2</a>
architecture for visual instruction tuning.</p>
<!-- raw HTML omitted -->
</blockquote>
<p>... (truncated)</p>
</details>
<details>
<summary>Commits</summary>
<ul>
<li><a
href="https://github.com/huggingface/transformers/commit/b7ee1e80b912c6cdd93b54dd77af061fde151d28"><code>b7ee1e8</code></a>
v4.42.3</li>
<li><a
href="https://github.com/huggingface/transformers/commit/da50b41a272724dba0821d643a44dfebb23e3786"><code>da50b41</code></a>
Gemma capping is a must for big models (<a
href="https://redirect.github.com/huggingface/transformers/issues/31698">#31698</a>)</li>
<li><a
href="https://github.com/huggingface/transformers/commit/086c74efdf98b4e64ac40863ce190144316873a5"><code>086c74e</code></a>
v4.42.2</li>
<li><a
href="https://github.com/huggingface/transformers/commit/869186743da66563d07ea8ea81ebfc88bbf5139c"><code>8691867</code></a>
Fix Gemma2 4d attention mask (<a
href="https://redirect.github.com/huggingface/transformers/issues/31674">#31674</a>)</li>
<li><a
href="https://github.com/huggingface/transformers/commit/7edc9931c52d9de22d2c569122dfc3d927fd280e"><code>7edc993</code></a>
don't zero out the attention_mask when using sliding window with flash
attent...</li>
<li><a
href="https://github.com/huggingface/transformers/commit/e3cb841ca8b8735193ce912e6cf42f413cefa87c"><code>e3cb841</code></a>
v4.42.1</li>
<li><a
href="https://github.com/huggingface/transformers/commit/b2455e5b811c91e51aca70d8d4f9a79934af7052"><code>b2455e5</code></a>
[HybridCache] Fix <code>get_seq_length</code> method (<a
href="https://redirect.github.com/huggingface/transformers/issues/31661">#31661</a>)</li>
<li><a
href="https://github.com/huggingface/transformers/commit/6c1d0b069de22d7ed8aa83f733c25045eea0585d"><code>6c1d0b0</code></a>
Release: v4.42.0</li>
<li><a
href="https://github.com/huggingface/transformers/commit/69b0f44b812425fbe83bdb0f00f610b7f69e476a"><code>69b0f44</code></a>
Add gemma 2 (<a
href="https://redirect.github.com/huggingface/transformers/issues/31659">#31659</a>)</li>
<li><a
href="https://github.com/huggingface/transformers/commit/be50a0338b9d7b76448fcc9c5046a78118a4d968"><code>be50a03</code></a>
change anchor_image_size None for compatibility (<a
href="https://redirect.github.com/huggingface/transformers/issues/31640">#31640</a>)</li>
<li>Additional commits viewable in <a
href="https://github.com/huggingface/transformers/compare/v4.41.2...v4.42.3">compare
view</a></li>
</ul>
</details>
<br />


[![Dependabot compatibility
score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=transformers&package-manager=pip&previous-version=4.41.2&new-version=4.42.3)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores)

Dependabot will resolve any conflicts with this PR as long as you don't
alter it yourself. You can also trigger a rebase manually by commenting
`@dependabot rebase`.

[//]: # (dependabot-automerge-start)
[//]: # (dependabot-automerge-end)

---

<details>
<summary>Dependabot commands and options</summary>
<br />

You can trigger Dependabot actions by commenting on this PR:
- `@dependabot rebase` will rebase this PR
- `@dependabot recreate` will recreate this PR, overwriting any edits
that have been made to it
- `@dependabot merge` will merge this PR after your CI passes on it
- `@dependabot squash and merge` will squash and merge this PR after
your CI passes on it
- `@dependabot cancel merge` will cancel a previously requested merge
and block automerging
- `@dependabot reopen` will reopen this PR if it is closed
- `@dependabot close` will close this PR and stop Dependabot recreating
it. You can achieve the same result by closing it manually
- `@dependabot show <dependency name> ignore conditions` will show all
of the ignore conditions of the specified dependency
- `@dependabot ignore this major version` will close this PR and stop
Dependabot creating any more for this major version (unless you reopen
the PR or upgrade to it yourself)
- `@dependabot ignore this minor version` will close this PR and stop
Dependabot creating any more for this minor version (unless you reopen
the PR or upgrade to it yourself)
- `@dependabot ignore this dependency` will close this PR and stop
Dependabot creating any more for this dependency (unless you reopen the
PR or upgrade to it yourself)


</details>

Signed-off-by: dependabot[bot] <support@github.com>
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
  • Loading branch information
dependabot[bot] authored Jul 1, 2024
1 parent 861d0c3 commit 6db1d91
Showing 1 changed file with 20 additions and 18 deletions.
38 changes: 20 additions & 18 deletions poetry.lock

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

0 comments on commit 6db1d91

Please sign in to comment.