diff --git a/VERSION_NUMBER b/VERSION_NUMBER index 27f9cd322bb91..a8fdfda1c7824 100644 --- a/VERSION_NUMBER +++ b/VERSION_NUMBER @@ -1 +1 @@ -1.8.0 +1.8.1 diff --git a/cmake/CMakeLists.txt b/cmake/CMakeLists.txt index 35822e21079d6..0a1187f6b372b 100644 --- a/cmake/CMakeLists.txt +++ b/cmake/CMakeLists.txt @@ -1391,8 +1391,11 @@ if (onnxruntime_USE_CUDA) endif() enable_language(CUDA) message( STATUS "CMAKE_CUDA_COMPILER_VERSION: ${CMAKE_CUDA_COMPILER_VERSION}") - - set(CMAKE_CUDA_STANDARD 14) + if (WIN32) + set(CMAKE_CUDA_STANDARD 17) + else() + set(CMAKE_CUDA_STANDARD 14) + endif() file(TO_CMAKE_PATH ${onnxruntime_CUDNN_HOME} onnxruntime_CUDNN_HOME) set(ONNXRUNTIME_CUDA_LIBRARIES ${CUDA_LIBRARIES}) diff --git a/docs/Versioning.md b/docs/Versioning.md index 439d5ab7bc453..42ece36e742ed 100644 --- a/docs/Versioning.md +++ b/docs/Versioning.md @@ -26,6 +26,7 @@ For more details on ONNX Release versions, see [this page](https://github.com/on | ONNX Runtime release version | ONNX release version | ONNX opset version | ONNX ML opset version | Supported ONNX IR version | [Windows ML Availability](https://docs.microsoft.com/en-us/windows/ai/windows-ml/release-notes/)| |------------------------------|--------------------|--------------------|----------------------|------------------|------------------| +| 1.8.1 | **1.8** down to 1.2 | 13 | 2 | 7 | Windows AI 1.8+ | | 1.8.0 | **1.8** down to 1.2 | 13 | 2 | 7 | Windows AI 1.8+ | | 1.7.0 | **1.8** down to 1.2 | 13 | 2 | 7 | Windows AI 1.7+ | | 1.6.0 | **1.8** down to 1.2 | 13 | 2 | 7 | Windows AI 1.6+ | diff --git a/docs/python/README.rst b/docs/python/README.rst index d1681619df165..157c945356c00 100644 --- a/docs/python/README.rst +++ b/docs/python/README.rst @@ -8,6 +8,11 @@ For more information on ONNX Runtime, please see `aka.ms/onnxruntime `_ or the `Github project `_. """ -__version__ = "1.8.0" +__version__ = "1.8.1" __author__ = "Microsoft" # we need to do device version validation (for example to check Cuda version for an onnxruntime-training package). diff --git a/onnxruntime/core/providers/cpu/tensor/upsample.h b/onnxruntime/core/providers/cpu/tensor/upsample.h index c2aea9374b09a..0b48cd6fed396 100644 --- a/onnxruntime/core/providers/cpu/tensor/upsample.h +++ b/onnxruntime/core/providers/cpu/tensor/upsample.h @@ -18,8 +18,8 @@ constexpr const char* UpsampleModeCubic = "cubic"; // is a 4x4 matrix const size_t CubicModeGridLength = 4; -using GetNearestPixelFunc = std::function; -using GetOriginalCoordinateFunc = std::function; +using GetNearestPixelFunc = int64_t(*)(float, bool); +using GetOriginalCoordinateFunc = float (*)(float, float, float, float, float, float); enum UpsampleMode { NN = 0, // nearest neighbour diff --git a/onnxruntime/core/providers/cuda/cuda_execution_provider.h b/onnxruntime/core/providers/cuda/cuda_execution_provider.h index c1c35284ee879..e8848e06f2004 100644 --- a/onnxruntime/core/providers/cuda/cuda_execution_provider.h +++ b/onnxruntime/core/providers/cuda/cuda_execution_provider.h @@ -178,7 +178,7 @@ class CUDAExecutionProvider : public IExecutionProvider { p.reset(); }); } - std::shared_ptr p{std::make_shared()}; + std::shared_ptr p = std::make_shared(); }; static const std::shared_ptr& PerThreadContextCache() { diff --git a/package/rpm/onnxruntime.spec b/package/rpm/onnxruntime.spec index f91ebb9c7ff24..405e57b3cbf8f 100644 --- a/package/rpm/onnxruntime.spec +++ b/package/rpm/onnxruntime.spec @@ -1,5 +1,5 @@ Name: onnxruntime -Version: 1.8.0 +Version: 1.8.1 Release: 1%{?dist} Summary: onnxruntime