-
Notifications
You must be signed in to change notification settings - Fork 652
[ONNXRuntime] Added builds with CUDA and TensorRT Execution Providers #4386
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Merged
giordano
merged 7 commits into
JuliaPackaging:master
from
stemann:stemann/onnxruntime_providers_cuda
Sep 12, 2022
Merged
Changes from all commits
Commits
Show all changes
7 commits
Select commit
Hold shift + click to select a range
32e1084
[ONNXRuntime] Added builds with CUDA and TensorRT Execution Providers
stemann d0e73a8
Added support for aarch64-linux-gnu - for Jetson 4.6, i.e. CUDA 10.2
stemann 13f81cc
Limited CUDA builds to a single CUDA version per major CUDA version
stemann 17c64ca
Enabled lazy artifacts
stemann 92fe008
Moved ONNXRuntime to O/ONNXRuntime
stemann 6de99a7
Added comment wrt. CUDA platform expansion wrt. C++ string ABI
stemann cfadc5c
Reverted to looping over CUDA versions
stemann File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
There are no files selected for viewing
This file was deleted.
Oops, something went wrong.
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
| @@ -0,0 +1,132 @@ | ||
| # Note that this script can accept some limited command-line arguments, run | ||
| # `julia build_tarballs.jl --help` to see a usage message. | ||
| using BinaryBuilder, Pkg | ||
|
|
||
| name = "ONNXRuntime" | ||
| version = v"1.10.0" | ||
|
|
||
| # Cf. https://onnxruntime.ai/docs/execution-providers/CUDA-ExecutionProvider.html#requirements | ||
| # Cf. https://onnxruntime.ai/docs/execution-providers/TensorRT-ExecutionProvider.html#requirements | ||
| cuda_versions = [ | ||
| # No CUDA 10.2, since pre-built x86_64 CUDA binaries are built for CUDA 11 | ||
| v"11.3", # Using 11.3, and not 11.4, to be compatible with TensorRT (JLL) v8.0.1 (which includes aarch64 support) | ||
| ] | ||
| cuda_aarch64_tag = "10.2" | ||
| cudnn_version = v"8.2.4" | ||
| tensorrt_version = v"8.0.1" | ||
|
|
||
| cudnn_compat = string(cudnn_version.major) | ||
| tensorrt_compat = string(tensorrt_version.major) | ||
|
|
||
| # Collection of sources required to complete build | ||
| sources = [ | ||
| GitSource("https://github.com/microsoft/onnxruntime.git", "0d9030e79888d1d5828730b254fedc53c7b640c1"), | ||
| ArchiveSource("https://github.com/microsoft/onnxruntime/releases/download/v$version/onnxruntime-win-x64-$version.zip", "a0c6db3cff65bd282f6ba4a57789e619c27e55203321aa08c023019fe9da50d7"; unpack_target="onnxruntime-x86_64-w64-mingw32"), | ||
| ArchiveSource("https://github.com/microsoft/onnxruntime/releases/download/v$version/onnxruntime-win-x86-$version.zip", "fd1680fa7248ec334efc2564086e9c5e0d6db78337b55ec32e7b666164bdb88c"; unpack_target="onnxruntime-i686-w64-mingw32"), | ||
| ArchiveSource("https://github.com/microsoft/onnxruntime/releases/download/v$version/onnxruntime-linux-x64-gpu-$version.tgz", "bc880ba8a572acf79d50dcd35ba6dd8e5fb708d03883959ef60efbc15f5cdcb6"; unpack_target="onnxruntime-x86_64-linux-gnu-cuda"), | ||
| ArchiveSource("https://github.com/microsoft/onnxruntime/releases/download/v$version/onnxruntime-win-x64-gpu-$version.zip", "0da11b8d953fad4ec75f87bb894f72dea511a3940cff2f4dad37451586d1ebbc"; unpack_target="onnxruntime-x86_64-w64-mingw32-cuda"), | ||
| # aarch64-linux-gnu binaries for NVIDIA Jetson from NVIDIA-managed Jetson Zoo: https://elinux.org/Jetson_Zoo#ONNX_Runtime | ||
| FileSource("https://nvidia.box.com/shared/static/jy7nqva7l88mq9i8bw3g3sklzf4kccn2.whl", "a608b7a4a4fc6ad5c90d6005edbfe0851847b991b08aafff4549bbbbdb938bf6"; filename = "onnxruntime-aarch64-linux-gnu-cuda.whl"), | ||
| ] | ||
|
|
||
| # Bash recipe for building across all platforms | ||
| script = raw""" | ||
| cd $WORKSPACE/srcdir | ||
|
|
||
| if [[ $bb_full_target != *cuda* | ||
| || $bb_full_target == aarch64-linux-gnu*cuda* ]]; then # pre-built aarch64-linux-gnu binaries needs separate libonnxruntime (for CPU). | ||
| if [[ $target == *-w64-mingw32* ]]; then | ||
| chmod 755 onnxruntime-$target/onnxruntime-*/lib/* | ||
| mkdir -p $includedir $libdir | ||
| cp -av onnxruntime-$target/onnxruntime-*/include/* $includedir | ||
| cp -av onnxruntime-$target/onnxruntime-*/lib/* $libdir | ||
| install_license onnxruntime-$target/onnxruntime-*/LICENSE | ||
| else | ||
| # Cross-compiling for aarch64-apple-darwin on x86_64 requires setting arch.: https://github.com/microsoft/onnxruntime/blob/v1.10.0/cmake/CMakeLists.txt#L186 | ||
| if [[ $target == aarch64-apple-darwin* ]]; then | ||
| cmake_extra_args="-DCMAKE_OSX_ARCHITECTURES='arm64'" | ||
| fi | ||
|
|
||
| cd onnxruntime | ||
| git submodule update --init --recursive | ||
| mkdir -p build | ||
| cd build | ||
| cmake $WORKSPACE/srcdir/onnxruntime/cmake \ | ||
| -DCMAKE_INSTALL_PREFIX=$prefix \ | ||
| -DCMAKE_TOOLCHAIN_FILE=${CMAKE_TARGET_TOOLCHAIN} \ | ||
| -DCMAKE_BUILD_TYPE=Release \ | ||
| -DONNX_CUSTOM_PROTOC_EXECUTABLE=$host_bindir/protoc \ | ||
| -Donnxruntime_BUILD_SHARED_LIB=ON \ | ||
| -Donnxruntime_BUILD_UNIT_TESTS=OFF \ | ||
| -Donnxruntime_DISABLE_RTTI=OFF \ | ||
| $cmake_extra_args | ||
| make -j $nproc | ||
| make install | ||
| install_license $WORKSPACE/srcdir/onnxruntime/LICENSE | ||
| fi | ||
| else | ||
| if [[ $target == *-w64-mingw32* ]]; then | ||
| chmod 755 onnxruntime-$target-cuda/onnxruntime-*/lib/* | ||
| fi | ||
| mkdir -p $includedir $libdir | ||
| cp -av onnxruntime-$target-cuda/onnxruntime-*/include/* $includedir | ||
| find onnxruntime-$target-cuda/onnxruntime-*/lib -not -type d | xargs -Isrc cp -av src $libdir | ||
| install_license onnxruntime-$target-cuda/onnxruntime-*/LICENSE | ||
| fi | ||
|
|
||
| if [[ $bb_full_target == aarch64-linux-gnu*cuda* ]]; then | ||
| cd $WORKSPACE/srcdir | ||
| unzip -d onnxruntime-$target-cuda onnxruntime-$target-cuda.whl | ||
| mkdir -p $libdir | ||
| find onnxruntime-$target-cuda/onnxruntime_gpu*.data/purelib/onnxruntime/capi -name *.so* -not -name *py* | xargs -Isrc cp -av src $libdir | ||
| fi | ||
| """ | ||
|
|
||
| # These are the platforms we will build for by default, unless further | ||
| # platforms are passed in on the command line | ||
| function platform_exclude_filter(p::Platform) | ||
| libc(p) == "musl" || | ||
| p == Platform("i686", "Linux") || # No binary - and source build fails linking CXX shared library libonnxruntime.so | ||
| Sys.isfreebsd(p) | ||
| end | ||
| platforms = supported_platforms(; exclude=platform_exclude_filter) | ||
| platforms = expand_cxxstring_abis(platforms; skip=!Sys.islinux) | ||
|
|
||
| cuda_platforms = Platform[] | ||
| for cuda_version in cuda_versions | ||
| cuda_tag = "$(cuda_version.major).$(cuda_version.minor)" | ||
| for p in [ | ||
| Platform("x86_64", "Linux"; cuda = cuda_tag), | ||
| Platform("x86_64", "Windows"; cuda = cuda_tag) | ||
| ] | ||
| push!(platforms, p) | ||
| push!(cuda_platforms, p) | ||
| end | ||
| end | ||
| push!(platforms, Platform("aarch64", "Linux"; cuda = cuda_aarch64_tag)) | ||
| push!(cuda_platforms, Platform("aarch64", "Linux"; cuda = cuda_aarch64_tag)) | ||
|
|
||
| # Not expanding cuda_platforms wrt. cxxstring ABIs, since pre-built x86_64 CUDA binaries are built for cxx03 (and the tensorrt provider for both x86_64 and aarch64 is built for cxx11) | ||
|
|
||
| # The products that we will ensure are always built | ||
| products = [ | ||
| LibraryProduct(["libonnxruntime", "onnxruntime"], :libonnxruntime) | ||
| ] | ||
|
|
||
| # Dependencies that must be installed before this package can be built | ||
| dependencies = [ | ||
| Dependency("CUDNN_jll", cudnn_version; | ||
| compat = cudnn_compat, | ||
| platforms = cuda_platforms), | ||
| Dependency("TensorRT_jll", tensorrt_version; | ||
| compat = tensorrt_compat, | ||
| platforms = cuda_platforms), | ||
| Dependency("Zlib_jll"; platforms = cuda_platforms), | ||
| HostBuildDependency(PackageSpec("protoc_jll", Base.UUID("c7845625-083e-5bbe-8504-b32d602b7110"), v"3.16.1")) | ||
| ] | ||
|
|
||
| # Build the tarballs, and possibly a `build.jl` as well. | ||
| build_tarballs(ARGS, name, version, sources, script, platforms, products, dependencies; | ||
| lazy_artifacts = true, | ||
| julia_compat = "1.6", | ||
| preferred_gcc_version = v"8") | ||
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Uh oh!
There was an error while loading. Please reload this page.