Skip to content

Conversation

dependabot[bot]
Copy link

@dependabot dependabot bot commented on behalf of github Sep 23, 2025

Updated Microsoft.ML.OnnxRuntime from 1.19.0 to 1.22.1.

Release notes

Sourced from Microsoft.ML.OnnxRuntime's releases.

1.22.1

What's new?

This release replaces static linking of dxcore.lib with optional runtime loading, lowering the minimum supported version from Windows 10 22H2 (10.0.22621) to 20H1 (10.0.19041). This enables compatibility with Windows Server 2019 (10.0.17763), where dxcore.dll may be absent.

  • change dependency from gitlab eigen to github eigen-mirror #​24884 - @​prathikr
  • Weaken dxcore dependency #​24845 - @​skottmckay
  • [DML] Restore compatibility with Windows Sdk 10.0.17134.0 #​24950 - @​JulienMaille
  • Disable VCPKG's binary cache #​24889 - @​snnn

1.22

Announcements

  • This release introduces new API's for Model Editor, Auto EP infrastructure, and AOT Compile
  • OnnxRuntime GPU packages require CUDA 12.x , packages built for CUDA 11.x are no longer published.
  • The min supported Windows version is now 10.0.19041.

GenAI & Advanced Model Features

  • Constrained Decoding: Introduced new capabilities for constrained decoding, offering more control over generative AI model outputs.

Execution & Core Optimizations

Core

  • Auto EP Selection Infrastructure: Added foundational infrastructure to enable automatic selection of Execution Providers via selection policies, aiming to simplify configuration and optimize performance. (Pull Request #​24430)
  • Compile API: Introduced new APIs to support explicit compilation of ONNX models.
  • Model Editor API api's for creating or editing ONNX models

Execution Provider (EP) Updates

CPU EP/MLAS

  • KleidiAI Integration: Integrated KleidiAI into ONNX Runtime/MLAS for enhanced performance on Arm architectures.
  • MatMulNBits Support: Added support for MatMulNBits, enabling matrix multiplication with weights quantized to 8 bits.
  • GroupQueryAttention optimizations and enhancements

OpenVINO EP

  • Added support up to OpenVINO 2025.1
  • Introduced Intel compiler level optimizations for QDQ models.
  • Added support to select Intel devices based on LUID
  • Load_config feature improvement to support AUTO, HETERO and MULTI plugin.
  • misc bugfixes/optimizations
  • For detailed updates, refer to Pull Request #​24394: ONNXRuntime OpenVINO - Release 1.22

QNN EP

  • SDK Update: Added support for QNN SDK 2.33.2.
  • operator updates/support to Sum, Softmax, Upsample, Expand, ScatterND, Einsum
  • QNN EP can be built as shared or static library.
  • enable QnnGpu backend
  • For detailed updates refer to recent QNN tagged PR's

TensorRT EP

  • TensorRT Version: Added support for TensorRT 10.9.
    • Note for onnx-tensorrt open-source parser users: Please check here for specific requirements (Referencing 1.21 link as a placeholder, this should be updated for 1.22).
  • New Features:
    • EP option to enable TRT Preview Feature
    • Support to load TensorRT V3 plugin
  • Bug Fixes:
    • Resolved an issue related to multithreading scenarios.
      ... (truncated)

1.21.1

What's new?

  • Extend CMAKE_CUDA_FLAGS with all Blackwell compute capacity #​23928 - @​yf711
  • [ARM CPU] Fix fp16 const initialization on no-fp16 platform #​23978 - @​fajin-corp
  • [TensorRT EP] Call cudaSetDevice at compute function for handling multithreading scenario #​24010 - @​chilo-ms
  • Fix attention bias broadcast #​24017 - @​tianleiwu
  • Deleted the constant SKIP_CUDA_TEST_WITH_DML #​24113 - @​CodingSeaotter
  • [QNN EP] ARM64EC python package remove --vcpkg in build #​24174 - @​jywu-msft
  • [wasm] remove --vcpkg in wasm build #​24179 - @​fs-eire

1.21.0

Announcements

  • No large announcements of note this release! We've made a lot of small refinements to streamline your ONNX Runtime experience.

GenAI & Advanced Model Features

Enhanced Decoding & Pipeline Support

  • Added "chat mode" support for CPU, GPU, and WebGPU.
  • Provided support for decoder model pipelines.
  • Added support for Java API for MultiLoRA.

API & Compatibility Updates

Bug Fixes for Model Output

  • Fixed Phi series garbage output issues with long prompts.
  • Resolved gibberish issues with top_k on CPU.

Execution & Core Optimizations

Core Refinements

  • Reduced default logger usage for improved efficiency(#​23030).
  • Fixed a visibility issue in theadpool (#​23098).

Execution Provider (EP) Updates

General

  • Removed TVM EP from the source tree(#​22827).
  • Marked NNAPI EP for deprecation (following Google's deprecation of NNAPI).
  • Fixed a DLL delay loading issue that impacts WebGPU EP and DirectML EP's usability on Windows (#​23111, #​23227)

TensorRT EP Improvements

  • Added support for TensorRT 10.8.
  • Assigned DDS ops (NMS, RoiAlign, NonZero) to TensorRT by default.
  • Introduced option trt_op_types_to_exclude to exclude specific ops from TensorRT assignment.

CUDA EP Improvements

QNN EP Improvements

  • Introduced QNN shared memory support.
  • Improved performance for AI Hub models.
  • Added support for QAIRT/QNN SDK 2.31.
  • Added Python 3.13 package.
  • Miscellaneous bug fixes and enhancements.
  • QNN EP is now built as a shared library/DLL by default. To retain previous build behavior, use build option --use_qnn static_lib.

DirectML EP Support & Upgrades

  • Updated DirectML version from 1.15.2 to 1.15.4(#​22635).

... (truncated)

1.20.2

What's new?

Build System & Packages

  • Merge Windows machine pools for Web CI pipeline to reduce maintenance costs (#​23243) - @​snnn
  • Update boost URL for React Native CI pipeline (#​23281) - @​jchen351
  • Move ORT Training pipeline to GitHub actions and enable CodeQL scan for the source code (#​22543) - @​snnn
  • Move Linux GitHub actions to a dedicated machine pool (#​22566) - @​snnn
  • Update Apple deployment target to iOS 15.1 and macOS 13.3 (#​23308) - @​snnn
  • Deprecate macOS 12 in packaging pipeline (#​23017) - @​mszhanyi
  • Remove net8.0-android MAUI target from MAUI test project (#​23607) - @​carzh

CUDA EP

  • Fixes use of numeric_limits that causes a compiler error in Visual Studio 2022 v17.12 Preview 5 (#​22738, #​22868) - @​tianleiwu

QNN EP

  • Enable offloading graph input quantization and graph output dequantization to CPU by default. Improves inference latency by reducing the amount of I/O data copied between CPU and NPU. (#​23368) - @​adrianlizarraga

1.20.1

What's new?

Python Quantization Tool

  • Prevent int32 quantized bias from clipping by adjusting the weight's scale (#​22020) - @​adrianlizarraga
  • Update QDQ Pad, Slice, Softmax (#​22676) - @​adrianlizarraga
  • Introduce get_qdq_config() helper to get QDQ configurations (#​22677) - @​adrianlizarraga
  • Add reduce_range option to get_qdq_config() (#​22782) - @​adrianlizarraga
  • Flaky test due to Pad reflect bug (#​22798) - @​adrianlizarraga

CPU EP

  • Refactor SkipLayerNorm implementation to address issues (#​22719, #​22862) - @​amarin16, @​liqunfu

QNN EP

  • Add QNN SDK v2.28.2 support (#​22724, #​22844) - @​HectorSVC, @​adrianlizarraga

TensorRT EP

  • Exclude DDS ops from running on TRT (#​22875) - @​chilo-ms

Packaging

  • Rework the native library usage so that a pre-built ORT native package can be easily used (#​22345) - @​skottmckay
  • Fix Maven Sha256 Checksum Issue (#​22600) - @​idiskyle

Contributions

Big thank you to the release manager @​yf711, along with @​adrianlizarraga, @​HectorSVC, @​jywu-msft, and everyone else who helped to make this patch release process a smooth one!

1.20.0

Release Manager: @​apsonawane

Announcements

  • All ONNX Runtime Training packages have been deprecated. ORT 1.19.2 was the last release for which onnxruntime-training (PyPI), onnxruntime-training-cpu (PyPI), Microsoft.ML.OnnxRuntime.Training (Nuget), onnxruntime-training-c (CocoaPods), onnxruntime-training-objc (CocoaPods), and onnxruntime-training-android (Maven Central) were published.
  • ONNX Runtime packages will stop supporting Python 3.8 and Python 3.9. This decision aligns with NumPy Python version support. To continue using ORT with Python 3.8 and Python 3.9, you can use ORT 1.19.2 and earlier.
  • ONNX Runtime 1.20 CUDA packages will include new dependencies that were not required in 1.19 packages. The following dependencies are new: libcudnn_adv.so.9, libcudnn_cnn.so.9, libcudnn_engines_precompiled.so.9, libcudnn_engines_runtime_compiled.so.9, libcudnn_graph.so.9, libcudnn_heuristic.so.9, libcudnn_ops.so.9, libnvrtc.so.12, and libz.so.1.

Build System & Packages

  • Python 3.13 support is included in PyPI packages.
  • ONNX 1.17 support will be delayed until a future release, but the ONNX version used by ONNX Runtime has been patched to include a shape inference change to the Einsum op.
  • DLLs in the Maven build are now digitally signed (fix for issue reported here).
  • (Experimental) vcpkg support added for the CPU EP. The DML EP does not yet support vcpkg, and other EPs have not been tested.

Core

  • MultiLoRA support.
  • Reduced memory utilization.
    • Fixed alignment that was causing mmap to fail for external weights.
    • Eliminated double allocations when deserializing external weights.
    • Added ability to serialize pre-packed weights so that they don’t cause an increase in memory utilization when the model is loaded.
  • Support bfloat16 and float8 data types in python I/O binding API.

Performance

  • INT4 quantized embedding support on CPU and CUDA EPs.
  • Miscellaneous performance improvements and bug fixes.

EPs

CPU

  • FP16 support for MatMulNbits, Clip, and LayerNormalization ops.

CUDA

  • Cudnn frontend integration for convolution operators.
  • Added support of cuDNN Flash Attention and Lean Attention in MultiHeadAttention op.

TensorRT

QNN

  • QNN HTP support for weight sharing across multiple ORT inference sessions. (See ORT QNN EP documentation for more information.)
  • Support for QNN SDK 2.27.

OpenVINO

  • Added support up to OpenVINO 2024.4.1.
  • Compile-time memory optimizations.
  • Enhancement of ORT EPContext Session option for optimized first inference latency.
  • Added remote tensors to ensure direct memory access for inferencing on NPU.

DirectML

... (truncated)

1.19.2

Announcements

  • ORT 1.19.2 is a small patch release, fixing some broken workflows and introducing bug fixes.

Build System & Packages

  • Fixed the signing of native DLLs.
  • Disabled absl symbolize in Windows Release build to avoid dependency on dbghelp.dll.

Training

  • Restored support for CUDA compute capability 7.0 and 7.5 with CUDA 12, and 6.0 and 6.1 with CUDA 11.
  • Several fixes for training CI pipelines.

Mobile

  • Fixed ArgMaxOpBuilder::AddToModelBuilderImpl() nullptr Node access for CoreML EP.

Generative AI

  • Added CUDA kernel for Phi3 MoE.
  • Added smooth softmax support in CUDA and CPU kernels for the GroupQueryAttention operator.
  • Fixed number of splits calculations in GroupQueryAttention CUDA operator.
  • Enabled causal support in the MultiHeadAttention CUDA operator.

Contributors

@​prathikr, @​mszhanyi, @​edgchen1, @​tianleiwu, @​wangyems, @​aciddelgado, @​mindest, @​snnn, @​baijumeswani, @​MaanavD

Thanks to everyone who helped ship this release smoothly!

Full Changelog: microsoft/onnxruntime@v1.19.0...v1.19.2

Commits viewable in compare view.

Dependabot compatibility score

Dependabot will resolve any conflicts with this PR as long as you don't alter it yourself. You can also trigger a rebase manually by commenting @dependabot rebase.


Dependabot commands and options

You can trigger Dependabot actions by commenting on this PR:

  • @dependabot rebase will rebase this PR
  • @dependabot recreate will recreate this PR, overwriting any edits that have been made to it
  • @dependabot merge will merge this PR after your CI passes on it
  • @dependabot squash and merge will squash and merge this PR after your CI passes on it
  • @dependabot cancel merge will cancel a previously requested merge and block automerging
  • @dependabot reopen will reopen this PR if it is closed
  • @dependabot close will close this PR and stop Dependabot recreating it. You can achieve the same result by closing it manually
  • @dependabot show <dependency name> ignore conditions will show all of the ignore conditions of the specified dependency
  • @dependabot ignore this major version will close this PR and stop Dependabot creating any more for this major version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this minor version will close this PR and stop Dependabot creating any more for this minor version (unless you reopen the PR or upgrade to it yourself)
  • @dependabot ignore this dependency will close this PR and stop Dependabot creating any more for this dependency (unless you reopen the PR or upgrade to it yourself)

---
updated-dependencies:
- dependency-name: Microsoft.ML.OnnxRuntime
  dependency-version: 1.22.1
  dependency-type: direct:production
  update-type: version-update:semver-minor
...

Signed-off-by: dependabot[bot] <support@github.com>
@dependabot @github
Copy link
Author

dependabot bot commented on behalf of github Sep 23, 2025

Labels

The following labels could not be found: dependencies, nuget. Please create them before Dependabot can add them to a pull request.

Please fix the above issues or remove invalid values from dependabot.yml.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

0 participants