Skip to content

Update dependency com.microsoft.onnxruntime:onnxruntime to v1.22.0 #79

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: main
Choose a base branch
from

Conversation

renovate[bot]
Copy link
Contributor

@renovate renovate bot commented Mar 8, 2025

This PR contains the following updates:

Package Change Age Adoption Passing Confidence
com.microsoft.onnxruntime:onnxruntime (source) 1.20.0 -> 1.22.0 age adoption passing confidence

Release Notes

microsoft/onnxruntime (com.microsoft.onnxruntime:onnxruntime)

v1.22.0: ONNX Runtime v1.22

Compare Source

Announcements

  • This release introduces new API's for Model Editor, Auto EP infrastructure, and AOT Compile
  • OnnxRuntime GPU packages require CUDA 12.x , packages built for CUDA 11.x are no longer published.

GenAI & Advanced Model Features

  • Constrained Decoding: Introduced new capabilities for constrained decoding, offering more control over generative AI model outputs.

Execution & Core Optimizations

Core
  • Auto EP Selection Infrastructure: Added foundational infrastructure to enable automatic selection of Execution Providers via selection policies, aiming to simplify configuration and optimize performance. (Pull Request #​24430)
  • Compile API: Introduced new APIs to support explicit compilation of ONNX models.
  • Model Editor API api's for creating or editing ONNX models
Execution Provider (EP) Updates
CPU EP/MLAS
  • KleidiAI Integration: Integrated KleidiAI into ONNX Runtime/MLAS for enhanced performance on Arm architectures.
  • MatMulNBits Support: Added support for MatMulNBits, enabling matrix multiplication with weights quantized to 8 bits.
  • GroupQueryAttention optimizations and enhancements
OpenVINO EP
  • Added support up to OpenVINO 2025.1
  • Introduced Intel compiler level optimizations for QDQ models.
  • Added support to select Intel devices based on LUID
  • Load_config feature improvement to support AUTO, HETERO and MULTI plugin.
  • misc bugfixes/optimizations
  • For detailed updates, refer to Pull Request #​24394: ONNXRuntime OpenVINO - Release 1.22
QNN EP
  • SDK Update: Added support for QNN SDK 2.33.2.
  • operator updates/support to Sum, Softmax, Upsample, Expand, ScatterND, Einsum
  • QNN EP can be built as shared or static library.
  • enable QnnGpu backend
  • For detailed updates refer to recent QNN tagged PR's
TensorRT EP
  • TensorRT Version: Added support for TensorRT 10.9.
    • Note for onnx-tensorrt open-source parser users: Please check here for specific requirements (Referencing 1.21 link as a placeholder, this should be updated for 1.22).
  • New Features:
    • EP option to enable TRT Preview Feature
    • Support to load TensorRT V3 plugin
  • Bug Fixes:
    • Resolved an issue related to multithreading scenarios.
    • Fixed incorrect GPU usage that affected both TensorRT EP and CUDA EP.
NV TensorRT RTX EP
  • New Execution Provider: Introduced a new Execution Provider specifically for Nvidia RTX GPUs, leveraging TensorRT for optimized performance.
CUDA EP
  • MatMulNBits Enhancement: Added support for 8-bit weight-only quantization in MatMulNBits.
  • Bug Fixes:
    • Fixed incorrect GPU usage (also mentioned under TensorRT EP).
VitisAI EP
  • Miscellaneous bug fixes and improvements.

Infrastructure & Build Improvements

Build System & Packages
  • QNN Nuget Package: The QNN Nuget package is now built as ARM64x.
Dependencies / Version Updates
  • CUDA Version Update: This release includes an update to the CUDA version. Users should consult the documentation for specific version requirements. CUDA 11 based GPU packages no longer published.

Web

  • WebGPU Expansion:
    • Added WebGPU support to the node.js package (Windows and macOS).
    • Enabled WebGPU when building from source for macOS, Linux, and Windows.

Mobile

  • No major updates of note this release.

Contributions

Contributors to ONNX Runtime include members across teams at Microsoft, along with our community members:

Yulong Wang, Jian Chen, Changming Sun, Satya Kumar Jandhyala, Hector Li, Prathik Rao, Adrian Lizarraga, Jiajia Qin, Scott McKay, Jie Chen, Tianlei Wu, Edward Chen, Wanming Lin, xhcao, vraspar, Dmitri Smirnov, Jing Fang, Yifan Li, Caroline Zhu, Jianhui Dai, Chi Lo, Guenther Schmuelling, Ryan Hill, Sushanth Rajasankar, Yi-Hong Lyu, Ankit Maheshkar, Artur Wojcik, Baiju Meswani, David Fan, Enrico Galli, Hans, Jambay Kinley, John Paul, Peishen Yan, Yateng Hong, amarin16, chuteng-quic, kunal-vaishnavi, quic-hungjuiw, Alessio Soldano, Andreas Hussing, Ashish Garg, Ashwath Shankarnarayan, Chengdong Liang, Clément Péron, Erick Muñoz, Fanchen Kong, George Wu, Haik Silm, Jagadish Krishnamoorthy, Justin Chu, Karim Vadsariya, Kevin Chen, Mark Schofield, Masaya, Kato, Michael Tyler, Nenad Banfic, Ningxin Hu, Praveen G, Preetha Veeramalai, Ranjit Ranjan, Seungtaek Kim, Ti-Tai Wang, Xiaofei Han, Yueqing Zhang, co63oc, derdeljan-msft, genmingz@AMD, jiangzhaoming, jing-bao, kuanyul-quic, liqun Fu, minfhong-quic, mingyue, quic-tirupath, quic-zhaoxul, saurabh, selenayang888, sfatimar, sheetalarkadam, virajwad, zz002, Ștefan Talpalaru

v1.21.1: ONNX Runtime v1.21.1

Compare Source

What's new?

v1.21.0: ONNX Runtime v1.21.0

Compare Source

Announcements

  • No large announcements of note this release! We've made a lot of small refinements to streamline your ONNX Runtime experience.

GenAI & Advanced Model Features

Enhanced Decoding & Pipeline Support
  • Added "chat mode" support for CPU, GPU, and WebGPU.
  • Provided support for decoder model pipelines.
  • Added support for Java API for MultiLoRA.
API & Compatibility Updates
Bug Fixes for Model Output
  • Fixed Phi series garbage output issues with long prompts.
  • Resolved gibberish issues with top_k on CPU.

Execution & Core Optimizations

Core Refinements
  • Reduced default logger usage for improved efficiency(#​23030).
  • Fixed a visibility issue in theadpool (#​23098).
Execution Provider (EP) Updates
General
  • Removed TVM EP from the source tree(#​22827).
  • Marked NNAPI EP for deprecation (following Google's deprecation of NNAPI).
  • Fixed a DLL delay loading issue that impacts WebGPU EP and DirectML EP's usability on Windows (#​23111, #​23227)
TensorRT EP Improvements
  • Added support for TensorRT 10.8.
  • Assigned DDS ops (NMS, RoiAlign, NonZero) to TensorRT by default.
  • Introduced option trt_op_types_to_exclude to exclude specific ops from TensorRT assignment.
CUDA EP Improvements
QNN EP Improvements
  • Introduced QNN shared memory support.
  • Improved performance for AI Hub models.
  • Added support for QAIRT/QNN SDK 2.31.
  • Added Python 3.13 package.
  • Miscellaneous bug fixes and enhancements.
  • QNN EP is now built as a shared library/DLL by default. To retain previous build behavior, use build option --use_qnn static_lib.
DirectML EP Support & Upgrades
  • Updated DirectML version from 1.15.2 to 1.15.4(#​22635).
OpenVINO EP Improvements
  • Introduced OpenVINO EP Weights Sharing feature.
  • Added support for various contrib Ops in OVEP:
    • SkipLayerNormalization, MatMulNBits, FusedGemm, FusedConv, EmbedLayerNormalization, BiasGelu, Attention, DynamicQuantizeMatMul, FusedMatMul, QuickGelu, SkipSimplifiedLayerNormalization
  • Miscellaneous bug fixes and improvements.
VitisAI EP Improvements
  • Miscellaneous bug fixes and improvements.

Mobile Platform Enhancements

CoreML Updates
  • Added support for caching generated CoreML models.

Extensions & Tokenizer Improvements

Expanded Tokenizer Support
  • Now supports more tokenizer models, including ChatGLM, Baichuan2, Phi-4, etc.
  • Added full Phi-4 pre/post-processing support for text, vision, and audio.
  • Introduced RegEx pattern loading from tokenizer.json.
Image Codec Enhancements
  • ImageCodec now links to native APIs if available; otherwise, falls back to built-in libraries.
Unified Tokenizer API
  • Introduced a new tokenizer op schema to unify the tokenizer codebase.
  • Added support for loading tokenizer data from a memory blob in the C API.

Infrastructure & Build Improvements

Runtime Requirements

All the prebuilt Windows packages now require VC++ Runtime version >= 14.40(instead of 14.38). If your VC++ runtime version is lower than that, you may see a crash when ONNX Runtime was initializing. See https://github.com/microsoft/STL/wiki/Changelog#vs-2022-1710 for more details.

Updated minimum iOS and Android SDK requirements to align with React Native 0.76:

  • iOS >= 15.1
  • Android API >= 24 (Android 7)

All macOS packages now require macOS version >= 13.3.

CMake File Changes

CMake Version: Increased the minimum required CMake version from 3.26 to 3.28. Added support for CMake 4.0.
Python Version: Increased the minimum required Python version from 3.8 to 3.10 for building ONNX Runtime from source.
Improved VCPKG support

Added the following cmake options for WebGPU EP

  • onnxruntime_USE_EXTERNAL_DAWN
  • onnxruntime_CUSTOM_DAWN_SRC_PATH
  • onnxruntime_BUILD_DAWN_MONOLITHIC_LIBRARY
  • onnxruntime_ENABLE_PIX_FOR_WEBGPU_EP
  • onnxruntime_ENABLE_DAWN_BACKEND_VULKAN
  • onnxruntime_ENABLE_DAWN_BACKEND_D3D12

Added cmake option onnxruntime_BUILD_QNN_EP_STATIC_LIB for building with QNN EP as a static library.
Removed cmake option onnxruntime_USE_PREINSTALLED_EIGEN.

Fixed a build issue with Visual Studio 2022 17.3 (#​23911)

Modernized Build Tools
  • Now using VCPKG for most package builds.
  • Upgraded Gradle from 7.x to 8.x.
  • Updated JDK from 11 to 17.
  • Enabled onnxruntime_USE_CUDA_NHWC_OPS by default for CUDA builds.
  • Added support for WASM64 (build from source; no package published).
Dependency Cleanup
  • Removed Google’s nsync from dependencies.
Others

Updated Node.js installation script to support network proxy usage (#​23231)

Web

  • No updates of note.

Contributors

Contributors to ONNX Runtime include members across teams at Microsoft, along with our community members:

Changming Sun, Yulong Wang, Tianlei Wu, Jian Chen, Wanming Lin, Adrian Lizarraga, Hector Li, Jiajia Qin, Yifan Li, Edward Chen, Prathik Rao, Jing Fang, shiyi, Vincent Wang, Yi Zhang, Dmitri Smirnov, Satya Kumar Jandhyala, Caroline Zhu, Chi Lo, Justin Chu, Scott McKay, Enrico Galli, Kyle, Ted Themistokleous, dtang317, wejoncy, Bin Miao, Jambay Kinley, Sushanth Rajasankar, Yueqing Zhang, amancini-N, ivberg, kunal-vaishnavi, liqun Fu, Corentin Maravat, Peishen Yan, Preetha Veeramalai, Ranjit Ranjan, Xavier Dupré, amarin16, jzm-intel, kailums, xhcao, A-Satti, Aleksei Nikiforov, Ankit Maheshkar, Javier Martinez, Jianhui Dai, Jie Chen, Jon Campbell, Karim Vadsariya, Michael Tyler, PARK DongHa, Patrice Vignola, Pranav Sharma, Sam Webster, Sophie Schoenmeyer, Ti-Tai Wang, Xu Xing, Yi-Hong Lyu, genmingz@AMD, junchao-zhao, sheetalarkadam, sushraja-msft, Akshay Sonawane, Alexis Tsogias, Ashrit Shetty, Bilyana Indzheva, Chen Feiyue, Christian Larson, David Fan, David Hotham, Dmitry Deshevoy, Frank Dong, Gavin Kinsey, George Wu, Grégoire, Guenther Schmuelling, Indy Zhu, Jean-Michaël Celerier, Jeff Daily, Joshua Lochner, Kee, Malik Shahzad Muzaffar, Matthieu Darbois, Michael Cho, Michael Sharp, Misha Chornyi, Po-Wei (Vincent), Sevag H, Takeshi Watanabe, Wu, Junze, Xiang Zhang, Xiaoyu, Xinpeng Dou, Xinya Zhang, Yang Gu, Yateng Hong, mindest, mingyue, raoanag, saurabh, shaoboyan091, sstamenk, tianf-fff, wonchung-microsoft, xieofxie, zz002


Configuration

📅 Schedule: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined).

🚦 Automerge: Disabled by config. Please merge this manually once you are satisfied.

Rebasing: Whenever PR is behind base branch, or you tick the rebase/retry checkbox.

🔕 Ignore: Close this PR and you won't be reminded about this update again.


  • If you want to rebase/retry this PR, check this box

This PR was generated by Mend Renovate. View the repository job log.

@renovate renovate bot force-pushed the renovate/com.microsoft.onnxruntime-onnxruntime-1.x branch 2 times, most recently from 57c44b9 to 29a32ac Compare March 20, 2025 18:18
@renovate renovate bot force-pushed the renovate/com.microsoft.onnxruntime-onnxruntime-1.x branch from 29a32ac to be006de Compare March 31, 2025 16:32
@renovate renovate bot force-pushed the renovate/com.microsoft.onnxruntime-onnxruntime-1.x branch 2 times, most recently from 74554fc to 08e5591 Compare April 15, 2025 17:10
@renovate renovate bot force-pushed the renovate/com.microsoft.onnxruntime-onnxruntime-1.x branch from 08e5591 to 76697d4 Compare April 21, 2025 10:51
@renovate renovate bot changed the title Update dependency com.microsoft.onnxruntime:onnxruntime to v1.21.0 Update dependency com.microsoft.onnxruntime:onnxruntime to v1.21.1 Apr 21, 2025
@renovate renovate bot changed the title Update dependency com.microsoft.onnxruntime:onnxruntime to v1.21.1 Update dependency com.microsoft.onnxruntime:onnxruntime to v1.22.0 May 10, 2025
@renovate renovate bot force-pushed the renovate/com.microsoft.onnxruntime-onnxruntime-1.x branch 2 times, most recently from b3cd8c1 to 58fbfb1 Compare May 16, 2025 09:18
@renovate renovate bot force-pushed the renovate/com.microsoft.onnxruntime-onnxruntime-1.x branch 6 times, most recently from 6928217 to f5098b3 Compare May 20, 2025 14:51
@renovate renovate bot force-pushed the renovate/com.microsoft.onnxruntime-onnxruntime-1.x branch from f5098b3 to b0a142d Compare May 20, 2025 14:54
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

0 participants