Skip to content

chore(deps): bump transformers from 4.48.3 to 4.53.2 in /backend/python/coqui #5851

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: master
Choose a base branch
from

Conversation

dependabot[bot]
Copy link
Contributor

@dependabot dependabot bot commented on behalf of github Jul 14, 2025

Bumps transformers from 4.48.3 to 4.53.2.

Release notes

Sourced from transformers's releases.

Patch Release v4.53.2

This patch contains the following bug fixes:

  • Fix some bug for finetune and batch infer For GLM-4.1V (#39090)
  • [bugfix] fix flash attention 2 unavailable error on Ascend NPU (#39166)
  • Fix errors when use verl to train GLM4.1v model (#39199)
  • [pagged-attention] fix off-by-1 error in pagged attention generation (#39258)
  • [smollm3] add tokenizer mapping for smollm3 (#39271)
  • [sliding window] revert and deprecate (#39301)
  • fix Glm4v batch videos forward (#39172)
  • Add a default value for position_ids in masking_utils (#39310)

Patch Release v4.53.1

This patch contains several bug fixes. The following commits are included:

  • Fix: unprotected import of tp plugin (#39083)
  • Fix key mapping for VLMs (#39029)
  • Several fixes for Gemma3n(#39135)
  • [qwen2-vl] fix FA2 inference (#39121)
  • [smolvlm] fix video inference (#39147)
  • Fix multimodal processor get duplicate arguments when receive kwargs for initialization (#39125)
  • when delaying optimizer creation only prepare the model (#39152)
  • Add packed tensor format support for flex/sdpa/eager through the mask! (#39194)

Release v4.53.0

Gemma3n

Gemma 3n models are designed for efficient execution on low-resource devices. They are capable of multimodal input, handling text, image, video, and audio input, and generating text outputs, with open weights for pre-trained and instruction-tuned variants. These models were trained with data in over 140 spoken languages.

Gemma 3n models use selective parameter activation technology to reduce resource requirements. This technique allows the models to operate at an effective size of 2B and 4B parameters, which is lower than the total number of parameters they contain. For more information on Gemma 3n's efficient parameter management technology, see the Gemma 3n page.

image

from transformers import pipeline
import torch
pipe = pipeline(
"image-text-to-text",
torch_dtype=torch.bfloat16,
model="google/gemma-3n-e4b",
device="cuda",
)
output = pipe(
"https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/bee.jpg",
text="<image_soft_token> in this image, there is"
)
print(output)
</tr></table>

... (truncated)

Commits

Dependabot compatibility score

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


Dependabot commands and options

You can trigger Dependabot actions by commenting on this PR:

  • @dependabot rebase will rebase this PR
  • @dependabot recreate will recreate this PR, overwriting any edits that have been made to it
  • @dependabot merge will merge this PR after your CI passes on it
  • @dependabot squash and merge will squash and merge this PR after your CI passes on it
  • @dependabot cancel merge will cancel a previously requested merge and block automerging
  • @dependabot reopen will reopen this PR if it is closed
  • @dependabot close will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
  • @dependabot show <dependency name> ignore conditions will show all of the ignore conditions of the specified dependency
  • @dependabot ignore this major version will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this minor version will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this dependency will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)

Bumps [transformers](https://github.com/huggingface/transformers) from 4.48.3 to 4.53.2.
- [Release notes](https://github.com/huggingface/transformers/releases)
- [Commits](huggingface/transformers@v4.48.3...v4.53.2)

---
updated-dependencies:
- dependency-name: transformers
  dependency-version: 4.53.2
  dependency-type: direct:production
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <[email protected]>
@github-actions github-actions bot enabled auto-merge (squash) July 14, 2025 21:51
Copy link

netlify bot commented Jul 14, 2025

Deploy Preview for localai ready!

Name Link
🔨 Latest commit caea73e
🔍 Latest deploy log https://app.netlify.com/projects/localai/deploys/68757bbf9922db000894d11f
😎 Deploy Preview https://deploy-preview-5851--localai.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
dependencies python Pull requests that update Python code
Projects
None yet
Development

Successfully merging this pull request may close these issues.

0 participants