Skip to content

Commit

Permalink
[onnxruntime] resurrecting test feature
Browse files Browse the repository at this point in the history
  • Loading branch information
luncliff committed May 25, 2024
1 parent e59bd8a commit a3c14cb
Show file tree
Hide file tree
Showing 2 changed files with 15 additions and 9 deletions.
5 changes: 1 addition & 4 deletions ports/onnxruntime/portfile.cmake
Original file line number Diff line number Diff line change
Expand Up @@ -117,10 +117,7 @@ if("cuda" IN_LIST FEATURES)
if(NOT DEFINED ENV{CUDA_PATH})
message(FATAL_ERROR "ENV{CUDA_PATH} is required. Please check the environment variable")
endif()
message(STATUS "Using CUDA: $ENV{CUDA_PATH}")
get_filename_component(CUDA_VERSION "$ENV{CUDA_PATH}" NAME)
string(REPLACE "v" "" CUDA_VERSION "${CUDA_VERSION}") # "v12.0" -> "12.0"
message(STATUS " version: ${CUDA_VERSION}")
message(STATUS "Using ENV{CUDA_PATH}: $ENV{CUDA_PATH}")
endif()

# see tools/ci_build/build.py
Expand Down
19 changes: 14 additions & 5 deletions ports/onnxruntime/vcpkg.json
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
{
"name": "onnxruntime",
"version-semver": "1.17.3",
"port-version": 1,
"port-version": 2,
"description": "ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator",
"homepage": "https://onnxruntime.ai/",
"license": "MIT",
Expand All @@ -24,7 +24,12 @@
"name": "nsync",
"platform": "!windows"
},
"onnx",
{
"name": "onnx",
"features": [
"disable-static-generation"
]
},
"optional-lite",
"protobuf",
{
Expand Down Expand Up @@ -74,9 +79,6 @@
"mpi"
]
},
"nccl": {
"description": "Build with NCCL support"
},
"openvino": {
"description": "Build with OpenVINO support",
"supports": "!(osx | ios | android | emscripten)",
Expand All @@ -101,6 +103,13 @@
}
]
},
"test": {
"description": "Build ONNXRuntime unit tests, micro-benchmarks",
"dependencies": [
"benchmark",
"gtest"
]
},
"xnnpack": {
"description": "Build with XNNPack support",
"dependencies": [
Expand Down

0 comments on commit a3c14cb

Please sign in to comment.