dependabot[bot] opened a new pull request, #1068:
URL: https://github.com/apache/singa/pull/1068

   Bumps [transformers](https://github.com/huggingface/transformers) from 2.5.1 
to 4.30.0.
   <details>
   <summary>Release notes</summary>
   <p><em>Sourced from <a 
href="https://github.com/huggingface/transformers/releases";>transformers's 
releases</a>.</em></p>
   <blockquote>
   <h2>v4.30.0: 100k, Agents improvements, Safetensors core dependency, 
Swiftformer, Autoformer, MobileViTv2, timm-as-a-backbone</h2>
   <h2>100k</h2>
   <p>Transformers has just reached 100k stars on GitHub, and to celebrate we 
wanted to highlight 100 projects in the vicinity of <code>transformers</code>  
and we have decided to create an <a 
href="https://github.com/huggingface/transformers/blob/main/awesome-transformers.md";>awesome-transformers</a>
 page to do just that.</p>
   <p>We accept PRs to add projects to the list!</p>
   <ul>
   <li>Top 100  by <a 
href="https://github.com/LysandreJik";><code>@​LysandreJik</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/22912";>#22912</a></li>
   <li>Add LlamaIndex to awesome-transformers.md  by <a 
href="https://github.com/ravi03071991";><code>@​ravi03071991</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/23484";>#23484</a></li>
   <li>add cleanlab to awesome-transformers tools list  by <a 
href="https://github.com/jwmueller";><code>@​jwmueller</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/23440";>#23440</a></li>
   </ul>
   <h2>4-bit quantization and QLoRA</h2>
   <p>By leveraging the <code>bitsandbytes</code> library by <a 
href="https://github.com/TimDettmers";><code>@​TimDettmers</code></a>, we add 
4-bit support to <code>transformers</code> models!</p>
   <ul>
   <li>4-bit QLoRA via bitsandbytes (4-bit base model + LoRA)  by <a 
href="https://github.com/TimDettmers";><code>@​TimDettmers</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/23479";>#23479</a></li>
   </ul>
   <h2>Agents</h2>
   <p>The Agents framework has been improved and continues to be stabilized. 
Among bug fixes, here are the important new features that were added:</p>
   <ul>
   <li>Local agent capabilities, to load a generative model directly from 
<code>transformers</code> instead of relying on APIs.</li>
   <li>Prompts are now hosted on the Hub, which means that anyone can fork the 
prompts and update them with theirs, to let other community contributors re-use 
them</li>
   <li>We add an <code>AzureOpenAiAgent</code> class to support Azure OpenAI 
agents.</li>
   </ul>
   <ul>
   <li>Add local agent  by <a 
href="https://github.com/sgugger";><code>@​sgugger</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/23438";>#23438</a></li>
   <li>Enable prompts on the Hub  by <a 
href="https://github.com/sgugger";><code>@​sgugger</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/23662";>#23662</a></li>
   <li>Add AzureOpenAiAgent  by <a 
href="https://github.com/sgugger";><code>@​sgugger</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/24058";>#24058</a></li>
   </ul>
   <h2>Safetensors</h2>
   <p>The <code>safetensors</code> library is a safe serialization framework 
for machine learning tensors. It has been audited and will become the default 
serialization framework for several organizations (Hugging Face, EleutherAI, 
Stability AI).</p>
   <p>It has now become a core dependency of <code>transformers</code>.</p>
   <ul>
   <li>Making <code>safetensors</code> a core dependency.  by <a 
href="https://github.com/Narsil";><code>@​Narsil</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/23254";>#23254</a></li>
   </ul>
   <h2>New models</h2>
   <h3>Swiftformer</h3>
   <p>The SwiftFormer paper introduces a novel efficient additive attention 
mechanism that effectively replaces the quadratic matrix multiplication 
operations in the self-attention computation with linear element-wise 
multiplications. A series of models called ‘SwiftFormer’ is built based on 
this, which achieves state-of-the-art performance in terms of both accuracy and 
mobile inference speed. Even their small variant achieves 78.5% top-1 
ImageNet1K accuracy with only 0.8 ms latency on iPhone 14, which is more 
accurate and 2× faster compared to MobileViT-v2.</p>
   <ul>
   <li>Add swiftformer  by <a 
href="https://github.com/shehanmunasinghe";><code>@​shehanmunasinghe</code></a> 
in <a 
href="https://redirect.github.com/huggingface/transformers/issues/22686";>#22686</a></li>
   </ul>
   <h3>Autoformer</h3>
   <p>This model augments the Transformer as a deep decomposition architecture, 
which can progressively decompose the trend and seasonal components during the 
forecasting process.</p>
   <ul>
   <li>[Time-Series] Autoformer model  by <a 
href="https://github.com/elisim";><code>@​elisim</code></a> in <a 
href="https://redirect.github.com/huggingface/transformers/issues/21891";>#21891</a></li>
   </ul>
   <!-- raw HTML omitted -->
   </blockquote>
   <p>... (truncated)</p>
   </details>
   <details>
   <summary>Commits</summary>
   <ul>
   <li><a 
href="https://github.com/huggingface/transformers/commit/fe861e578f50dc9c06de33cd361d2f625017e624";><code>fe861e5</code></a>
 [<code>GPT2</code>] Add correct keys on 
<code>_keys_to_ignore_on_load_unexpected</code> on all chil...</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/b3e27a80578d022301611363b890107244e12354";><code>b3e27a8</code></a>
 Update the pin on Accelerate (<a 
href="https://redirect.github.com/huggingface/transformers/issues/24110";>#24110</a>)</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/53e1f5cf66d320b9c809f3940c707b6fef435d2d";><code>53e1f5c</code></a>
 [<code>Trainer</code>] Correct behavior of <code>_load_best_model</code> for 
PEFT models (<a 
href="https://redirect.github.com/huggingface/transformers/issues/24103";>#24103</a>)</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/17db177714b03103bb94cd71b7dd414bc63bffd5";><code>17db177</code></a>
 reset accelerate env variables after each test (<a 
href="https://redirect.github.com/huggingface/transformers/issues/24107";>#24107</a>)</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/905892f09027cab690918c7766fea1bb51bcdd26";><code>905892f</code></a>
 Release: v4.30.0</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/c3572e6bfba13ce6dc3fedb05cd1a946ea109576";><code>c3572e6</code></a>
 Add AzureOpenAiAgent (<a 
href="https://redirect.github.com/huggingface/transformers/issues/24058";>#24058</a>)</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/5eb3d3c7023ed0522d3c743ee2e13d896a3aa788";><code>5eb3d3c</code></a>
 Up pinned accelerate version (<a 
href="https://redirect.github.com/huggingface/transformers/issues/24089";>#24089</a>)</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/d1c039e39864a41f6eb8b770a65f123c40164ea5";><code>d1c039e</code></a>
 fix accelerator prepare during eval only mode (<a 
href="https://redirect.github.com/huggingface/transformers/issues/24014";>#24014</a>)</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/2c887cf8e0cb1ac96d28361ff3235a77f83c36ee";><code>2c887cf</code></a>
 Do not prepare lr scheduler as it as the right number of steps (<a 
href="https://redirect.github.com/huggingface/transformers/issues/24088";>#24088</a>)</li>
   <li><a 
href="https://github.com/huggingface/transformers/commit/12298cb65c7e9d615b749dde935a0b4966f4ae49";><code>12298cb</code></a>
 fix executable batch size issue (<a 
href="https://redirect.github.com/huggingface/transformers/issues/24067";>#24067</a>)</li>
   <li>Additional commits viewable in <a 
href="https://github.com/huggingface/transformers/compare/v2.5.1...v4.30.0";>compare
 view</a></li>
   </ul>
   </details>
   <br />
   
   
   [![Dependabot compatibility 
score](https://dependabot-badges.githubapp.com/badges/compatibility_score?dependency-name=transformers&package-manager=pip&previous-version=2.5.1&new-version=4.30.0)](https://docs.github.com/en/github/managing-security-vulnerabilities/about-dependabot-security-updates#about-compatibility-scores)
   
   Dependabot will resolve any conflicts with this PR as long as you don't 
alter it yourself. You can also trigger a rebase manually by commenting 
`@dependabot rebase`.
   
   [//]: # (dependabot-automerge-start)
   [//]: # (dependabot-automerge-end)
   
   ---
   
   <details>
   <summary>Dependabot commands and options</summary>
   <br />
   
   You can trigger Dependabot actions by commenting on this PR:
   - `@dependabot rebase` will rebase this PR
   - `@dependabot recreate` will recreate this PR, overwriting any edits that 
have been made to it
   - `@dependabot merge` will merge this PR after your CI passes on it
   - `@dependabot squash and merge` will squash and merge this PR after your CI 
passes on it
   - `@dependabot cancel merge` will cancel a previously requested merge and 
block automerging
   - `@dependabot reopen` will reopen this PR if it is closed
   - `@dependabot close` will close this PR and stop Dependabot recreating it. 
You can achieve the same result by closing it manually
   - `@dependabot ignore this major version` will close this PR and stop 
Dependabot creating any more for this major version (unless you reopen the PR 
or upgrade to it yourself)
   - `@dependabot ignore this minor version` will close this PR and stop 
Dependabot creating any more for this minor version (unless you reopen the PR 
or upgrade to it yourself)
   - `@dependabot ignore this dependency` will close this PR and stop 
Dependabot creating any more for this dependency (unless you reopen the PR or 
upgrade to it yourself)
   You can disable automated security fix PRs for this repo from the [Security 
Alerts page](https://github.com/apache/singa/network/alerts).
   
   </details>


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]

Reply via email to