onnxruntime icon indicating copy to clipboard operation
onnxruntime copied to clipboard

ORT 1.17.0 Release Candidates available for testing

Open YUNQIUGUO opened this issue 5 months ago • 19 comments

ORT 1.17 will be released in late January. Release candidate builds are available now for testing. If you encounter issues, please report them by responding in this issue.

Release branch: rel-1.17.0 Release manager: @YUNQIUGUO

Pypi Nuget npm Maven (Java)
CPU: 1.17.0.dev20240118001
GPU: 1.17.0.dev20240118001
CPU: 1.17.0-dev-20240119-0139-a63b71eadb
GPU (CUDA/TRT): 1.17.0-dev-20240118-2301-a63b71eadb
DirectML: 1.17.0-dev-20240119-0131-a63b71eadb
WindowsAI: 1.17.0-dev-20240119-0131-a63b71eadb
onnxruntime-node: 1.17.0-dev.20240118-a63b71eadb
onnxruntime-react-native 1.17.0-dev.20240118-a63b71eadb
onnxruntime-web 1.17.0-dev.20240118-a63b71eadb
CPU: 1.17.0-rc1
GPU: 1.17.0-rc1

Describe scenario use case

not applicable.

YUNQIUGUO avatar Jan 23 '24 06:01 YUNQIUGUO

@martinb35 / @smk2007

fdwr avatar Jan 23 '24 06:01 fdwr

Please share CUDA 12 packages for Python and Nuget.

tianleiwu avatar Jan 23 '24 21:01 tianleiwu

Agree, prebuilt CUDA 12.1 packages would be really appreciated. Like you started doing with this nightly: https://aiinfra.visualstudio.com/PublicPackages/_artifacts/feed/ort-cuda-12-nightly/PyPI/ort-nightly-gpu/versions/1.17.0.dev20231205004

dbuades avatar Jan 24 '24 18:01 dbuades

Here's the CUDA 12 version RC packages for Python and Nuget: (based on the same commit id)

ort-nightly-gpu : 1.17.0.dev20240118002

Microsoft.ML.OnnxRuntime.Managed:1.17.0-dev-20240118-2235-a63b71eadb

Microsoft.ML.OnnxRuntime.Gpu: 1.17.0-dev-20240118-2235-a63b71eadb

Microsoft.ML.OnnxRuntime.Gpu.Linux:1.17.0-dev-20240118-2235-a63b71eadb

Microsoft.ML.OnnxRuntime.Gpu.Windows:1.17.0-dev-20240118-2235-a63b71eadb

YUNQIUGUO avatar Jan 25 '24 19:01 YUNQIUGUO

@HectorSVC why was this issue unpinned? We pin our release candidates in GitHub issues so our partners + community members have easier access and can test them.

sophies927 avatar Feb 01 '24 22:02 sophies927

Thanks for managing the release @YUNQIUGUO. Given a PyPI release has been made for 1.17, is there a plan to tag and release it on GitHub as well? For context, the onnxruntime feedstock on conda-forge typically uses the tagged github release to build ORT.

adityagoel4512 avatar Feb 02 '24 00:02 adityagoel4512

Thanks for managing the release @YUNQIUGUO. Given a PyPI release has been made for 1.17, is there a plan to tag and release it on GitHub as well?

yep, we are still waiting for the last couple packages to be uploaded to the package management repo. after everything completed, will do a release announcement with the 1.17.0 package assets on Github as well.

YUNQIUGUO avatar Feb 02 '24 01:02 YUNQIUGUO

Thanks for the new release! However, I see that a CUDA 12 version of the python onnxruntime-gpu package wasn't included in the pypy release. I tested the CUDA 12 RC packages that @YUNQIUGUO published in this thread last week and everything was working well, so I believe releasing a 1.17 version would be very useful.

Do you have plans to upload them at a later date or is there a particular reason why you are choosing not to do it?

Thank you very much!

dbuades avatar Feb 02 '24 03:02 dbuades

The cuda 12 instructions are here: https://onnxruntime.ai/docs/install/#install-onnx-runtime-gpu-cuda-12x It seems the onnxruntime-gpu package doesn't exist on AzDo, I only found ort-nightly-gpu==1.17.0.dev20240130002 so far. Could we release a properly named artifact for 1.17, as opposed to the nightly build?

For pypi, i'm not sure pypi can release onnxruntime-gpu for more than one major version of cuda. E.g. pytorch hosts a separate repo for its cuda11 py packages.

fvdnabee avatar Feb 02 '24 15:02 fvdnabee

@YUNQIUGUO, please upload onnxruntime-gpu 1.17.0 CUDA 12 python package to https://aiinfra.visualstudio.com/PublicPackages/_artifacts/feed/onnxruntime-cuda-12/

tianleiwu avatar Feb 02 '24 17:02 tianleiwu

You are right @fvdnabee , thank you for reformulating my request.

dbuades avatar Feb 02 '24 17:02 dbuades

@fvdnabee

The cuda 12 instructions are here: https://onnxruntime.ai/docs/install/#install-onnx-runtime-gpu-cuda-12x It seems the onnxruntime-gpu package doesn't exist on AzDo, I only found ort-nightly-gpu==1.17.0.dev20240130002 so far. Could we release a properly named artifact for 1.17, as opposed to the nightly build?

Thanks for reporting this. I am guessing our Python Cuda 12 packaging pipeline probably lacks a release version configuration/nightly build option. I actually uploaded onnxruntime-gpu to the Cuda 12 official feed but looks like it still contains a nightly in the naming though it should be an official one instead. I will contact the pipeline owner and we'll look into address this issue and re-upload an official named 1.17.0 package. Since it's the first time release cuda 12 version wheels to the feed, sorry about that - the issue hasn't be identified before.

@dbuades I am not aware of a plan to upload it to an official repo like pypi.org yet. will ask around.

YUNQIUGUO avatar Feb 02 '24 17:02 YUNQIUGUO

I'm curious if onnxruntime-node now supports dml and cuda?

0x0480 avatar Feb 02 '24 20:02 0x0480

@fs-eire

YUNQIUGUO avatar Feb 02 '24 20:02 YUNQIUGUO

No. DML is ongoing (https://github.com/microsoft/onnxruntime/pull/19274) and CUDA support is the next.

I'm curious if onnxruntime-node now supports dml and cuda?

fs-eire avatar Feb 02 '24 21:02 fs-eire

@dbuades @fvdnabee Hey, the issue has been resolved now. and here's the official onnxruntime-gpu package for cuda 12 version: https://aiinfra.visualstudio.com/PublicPackages/_artifacts/feed/onnxruntime-cuda-12/PyPI/onnxruntime-gpu/overview/1.17.0

YUNQIUGUO avatar Feb 03 '24 00:02 YUNQIUGUO

No. DML is ongoing (#19274) and CUDA support is the next.

I'm curious if onnxruntime-node now supports dml and cuda?

#19274 is merged in main and marked for patch release 1.17.1, pending approval.

fs-eire avatar Feb 16 '24 01:02 fs-eire

I'm curious if there is any work towards DML for onnxruntime-java? Where might I be pointed to make that a thing? (I did search issues... seems not discussed yet?)

CaelumF avatar Feb 21 '24 14:02 CaelumF

I'm curious if there is any work towards DML for onnxruntime-java? Where might I be pointed to make that a thing? (I did search issues... seems not discussed yet?)

@CaelumF 🤔 I haven't heard desire for that combination of language + backend before (more often DML is accessed via C++, Python, and C#), but is it already supported? I see this enum ai.onnxruntime.OrtProvider.DIRECT_ML. (alas I don't know who on ORT owns the Java language layer to verify)

fdwr avatar Feb 22 '24 16:02 fdwr

@fdwr hey thanks for responding. The pre built jars do have functions for enabling DirectML, but the binary wasn't compiled with DirectML enabled and errors when attempting to use it, building with it enabled has some trouble. Issue posted here https://github.com/microsoft/onnxruntime/issues/19656 which also includes my use case in case you're curious

CaelumF avatar Feb 26 '24 21:02 CaelumF

I'm the maintainer of the Java layer, I'll have a look at what's going on.

Craigacp avatar Feb 26 '24 22:02 Craigacp