chore(deps): update transformers dep #627
Open
+5
−5
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Coming soon: The Renovate bot (GitHub App) will be renamed to Mend. PRs from Renovate will soon appear from 'Mend'. Learn more here.
This PR contains the following updates:
==1.10.0
->==1.10.1
==5.1.0
->==5.1.1
==4.55.4
->==4.56.2
Release Notes
huggingface/accelerate (accelerate)
v1.10.1
: : PatchfixCompare Source
Full Changelog: huggingface/accelerate@v1.10.0...v1.10.1
UKPLab/sentence-transformers (sentence-transformers)
v5.1.1
: - Explicit incorrect arguments, fixes for multi-GPU, evaluator, and hard negativeCompare Source
This patch makes Sentence Transformers more explicit with incorrect arguments and introduces some fixes for multi-GPU processing, evaluators, and hard negatives mining.
Install this version with
Error if unused kwargs is passed &
get_model_kwargs
(#3500)Some SentenceTransformer or SparseEncoder models support custom model-specific keyword arguments, such as jinaai/jina-embeddings-v4. As of this release, calling
model.encode
with keyword arguments that aren't used by the model will result in an error.Quite useful when you, for example, accidentally forget that the parameter to get normalized embeddings is
normalize_embeddings
. Prior to this version, this parameter would simply quietly be ignored.To check which custom extra keyword arguments may be used for your model, you can call the new
get_model_kwargs
method:Note: You can always pass the
task
parameter, it's the only model-specific parameter that will be quietly ignored. This means that you can always usemodel.encode(..., task="query")
andmodel.encode(..., task="document")
.Minor Features
Minor Fixes
batch_size
being ignored inCrossEncoderRerankingEvaluator
(#3497)encode
, embeddings are now moved from the various devices to the CPU before being stacked into one tensor (#3488)encode_query
andencode_document
inmine_hard_negatives
, automatically using defined "query" and "document" prompts (#3502)output_path
that doesn't exist yet (#3516)mine_hard_negatives
(#3504)All Changes
fix
] add batch size parameter to model prediction in CrossEncoderRerankingEvaluator by @emapco in #3497fix
] Ensure multi-process embeddings are moved to CPU for concatenation by @tomaarsen in #3488model_card
] Don't override manually provided languages in model card by @tomaarsen in #3501tests
] Add hard negatives test showing multiple positives are correctly handled by @tomaarsen in #3503feat
] Use encode_document and encode_query in mine_hard_negatives by @tomaarsen in #3502input_ids
,attention_mask
,token_type_ids
,inputs_embeds
to forward by @Samoed in #3509feat
] add get_model_kwargs method; throw error if unused kwarg is passed by @tomaarsen in #3500fix
] Fix the number of missing negatives in mine_hard_negatives by @tomaarsen in #3504New Contributors
Full Changelog: UKPLab/sentence-transformers@v5.1.0...v5.1.1
huggingface/transformers (transformers)
v4.56.2
: Patch release v4.56.2Compare Source
v4.56.1
: Patch release v4.56.1Compare Source
Patch release v4.56.1
This patch most notably fixes an issue with the new
dtype
argument (replacingtorch_dtype
) in pipelines!Bug Fixes & Improvements
v4.56.0
: v4.56: Dino v3, X-Codec, Ovis 2, MetaCLIP 2, Florence 2, SAM 2, Kosmos 2.5, HunYuan, GLMV-4.5Compare Source
New model additions
Dino v3
DINOv3 is a family of versatile vision foundation models that outperforms the specialized state of the art across a broad range of settings, without fine-tuning. DINOv3 produces high-quality dense features that achieve outstanding performance on various vision tasks, significantly surpassing previous self- and weakly-supervised foundation models.
You can find all the original DINOv3 checkpoints under the DINOv3 collection.
X-Codec
he X-Codec model was proposed in Codec Does Matter: Exploring the Semantic Shortcoming of Codec for Audio Language Model by Zhen Ye, Peiwen Sun, Jiahe Lei, Hongzhan Lin, Xu Tan, Zheqi Dai, Qiuqiang Kong, Jianyi Chen, Jiahao Pan, Qifeng Liu, Yike Guo, Wei Xue
The X-Codec model is a neural audio codec that integrates semantic information from self-supervised models (e.g., HuBERT) alongside traditional acoustic information. This enables :
Ovis 2
The Ovis2 is an updated version of the Ovis model developed by the AIDC-AI team at Alibaba International Digital Commerce Group.
Ovis2 is the latest advancement in multi-modal large language models (MLLMs), succeeding Ovis1.6. It retains the architectural design of the Ovis series, which focuses on aligning visual and textual embeddings, and introduces major improvements in data curation and training methods.
MetaCLIP 2
MetaCLIP 2 is a replication of the original CLIP model trained on 300+ languages. It achieves state-of-the-art (SOTA) results on multilingual benchmarks (e.g., XM3600, CVQA, Babel‑ImageNet), surpassing previous SOTA such as mSigLIP and SigLIP‑2. The authors show that English and non-English worlds can mutually benefit and elevate each other.
Florence 2
Florence-2 is an advanced vision foundation model that uses a prompt-based approach to handle a wide range of vision and vision-language tasks. Florence-2 can interpret simple text prompts to perform tasks like captioning, object detection, and segmentation. It leverages the FLD-5B dataset, containing 5.4 billion annotations across 126 million images, to master multi-task learning. The model's sequence-to-sequence architecture enables it to excel in both zero-shot and fine-tuned settings, proving to be a competitive vision foundation model.
SAM 2
SAM2 (Segment Anything Model 2) was proposed in Segment Anything in Images and Videos by Nikhila Ravi, Valentin Gabeur, Yuan-Ting Hu, Ronghang Hu, Chaitanya Ryali, Tengyu Ma, Haitham Khedr, Roman Rädle, Chloe Rolland, Laura Gustafson, Eric Mintun, Junting Pan, Kalyan Vasudev Alwala, Nicolas Carion, Chao-Yuan Wu, Ross Girshick, Piotr Dollár, Christoph Feichtenhofer.
The model can be used to predict segmentation masks of any object of interest given an input image or video, and input points or bounding boxes.
Kosmos 2.5
The Kosmos-2.5 model was proposed in KOSMOS-2.5: A Multimodal Literate Model by Microsoft.
The abstract from the paper is the following:
We present Kosmos-2.5, a multimodal literate model for machine reading of text-intensive images. Pre-trained on large-scale text-intensive images, Kosmos-2.5 excels in two distinct yet cooperative transcription tasks: (1) generating spatially-aware text blocks, where each block of text is assigned its spatial coordinates within the image, and (2) producing structured text output that captures styles and structures into the markdown format. This unified multimodal literate capability is achieved through a shared Transformer architecture, task-specific prompts, and flexible text representations. We evaluate Kosmos-2.5 on end-to-end document-level text recognition and image-to-markdown text generation. Furthermore, the model can be readily adapted for any text-intensive image understanding task with different prompts through supervised fine-tuning, making it a general-purpose tool for real-world applications involving text-rich images. This work also paves the way for the future scaling of multimodal large language models.
HunYuan
More information at release 🤗
Seed OSS
More information at release 🤗
GLM-4.5V
More information at release 🤗
Cache
Beyond a large refactor of the caching system in Transformers, making it much more practical and general, models using sliding window attention/chunk attention do not waste memory anymore when caching past states. It was allowed most notable by:
See the following improvements on memory usage for Mistral (using only sliding layers) and GPT-OSS (1 out of 2 layers is sliding) respectively:

Beyond memory usage, it will also improve generation/forward speed by a large margin for large contexts, as only necessary states are passed to the attention computation, which is very sensitive to the sequence length.
Quantization
MXFP4
Since the GPT-OSS release which introduced the MXPF4 quantization type, several improvements have been made to the support, which should now stabilize.
swiglu_limit
not passed in for MXFP4 by @danielhanchen in #40197Mxfp4
] Add a way to save with a quantization method by @ArthurZucker in #40176New standard
Now that we deprecated tensorflow and jax, we felt that
torch_dtype
was not only misaligned with torch, but was redundant and hard to remember. For this reason, we switched to a much more standarddtype
argument!torch_dtype
will still be a valid usage for as long as needed to ensure a smooth transition, but new code should usedtype
, and we encourage you to update older code as well!Breaking changes
The following commits are breaking changes in workflows that were either buggy or not working as expected.
Saner hub-defaults for hybrid cache implementation
On models where the hub checkpoint specifies cache_implementation="hybrid" (static sliding window hybrid cache), UNSETS this value. This will make the model use the dynamic sliding window layers by default.
This default meant that there were widespread super slow 1st generate calls on models with hybrid caches, which should nol onger be the case.
cache_implementation="hybrid"
hub defaults by @gante in #40135Sine positional embeddings for MaskFormer & LRU cache
Cache the computation of sine positional embeddings for MaskFormer; results in a 6% performance improvement.
Explicit cache initialization
Adds explicit cache initialization to prepare for the deprecation of the
from_legacy_cache
utility.Default compilation with
fullgraph=False
Having
fullgraph
set toTrue
during compilation ended up being very restrictive, especially with the arrival of widely-used MoEs.Remove decoding strategies
The DoLa decoding strategy has been moved to the following remote-code repository a few versions ago: https://huggingface.co/transformers-community/dola
The Contrastive Search decoding strategy has been moved to the following remote-code repository a few versions ago: https://huggingface.co/transformers-community/contrastive-search
Both have now been removed from the library as a result.
Fix sliding window in flash attention
Flash attention has used sliding window sizes which were off by one. This affected generations that had initially bigger contexts than the sliding window size.
Flash Attention
] Fix sliding window size by @vasqu in #40163Minimum Torch version is now 2.2
Torch 2.1 support has been unreliable for some time, so we've now made it official and bumped our minimum version to 2.2.
Bugfixes and improvements
GptOss
fixes for green CI by @gante in #39929utils/check_bad_commit.py
failing due to rate limit (requestingapi.github.com
) by @ydshieh in #39918torch.device('cpu').index
being None by @manueldeprada in #39933torchcodec
is updated by @ydshieh in #39951triton_kernels
dep withkernels
instead by @SunMarc in #39926fix_and_overwrite
mode ofutils/check_docstring.py
by @manueldeprada in #39369find_file_type
by @yonigozlan in #39897past_key_value
topast_key_valueS
everywhere by @Cyrilvallez in #39956notification_service.py
abouttime_spent
by @ydshieh in #40037notification_service.py
abouttime_spent
" by @ydshieh in #40044torchcodec==0.5.0
and usetorch 2.8
on daily CI by @ydshieh in #40072time_spent
innotification_service.py
. by @ydshieh in #40081GPT Big Code
] Fix attention scaling by @vasqu in #40041ForConditionalGeneration
by @qgallouedec in #39973is_fast
to ImageProcessor by @MilkClouds in #39603logger.warning
withlogger.warning_once
inGradientCheckpointingLayer
by @qgallouedec in #40091Flash Attention
] Fix flash attention integration by @vasqu in #40002custom_generate
collections by @gante in #39894tiny_agents.md
to Korean by @AhnJoonSung in #39913content
inputs for LLMs by @gante in #39829decoding_method
argument in generate by @manueldeprada in #40085generation_config
by @gante in #40127main_classes/processors.md
to Korean by @TaskerJang in #39519jamba.md
to Korean by @skwh54 in #39890main_classes/optimizer_schedules.md
to Korean by @luckyvickyricky in #39713gpt2.md
to Korean by @taemincode in #39808optimizers.md
to Korean by @chelsseeey in #40011pipelines.md
to Korean by @xhaktm00 in #39577gemma3.md
to Korean by @seopp in #39865torch_compile_test
andtorch_export_test
by @ydshieh in #39950self.tokenizer
byself.processing_class
by @qgallouedec in #40119too long with no output
by @ydshieh in #40201model_input_names
for PixtralImageProcessor by @rohitrango in #40226chat_template
(jinja2
) as an extra dependency by @tboerstad in #40128CI
] Fix repo consistency by @vasqu in #40249Configuration
📅 Schedule: Branch creation - "on friday" (UTC), Automerge - At any time (no schedule defined).
🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.
♻ Rebasing: Whenever PR is behind base branch, or you tick the rebase/retry checkbox.
👻 Immortal: This PR will be recreated if closed unmerged. Get config help if that's undesired.
This PR was generated by Mend Renovate. View the repository job log.