diff --git a/VERSION_NUMBER b/VERSION_NUMBER
index 27f9cd322bb91..a8fdfda1c7824 100644
--- a/VERSION_NUMBER
+++ b/VERSION_NUMBER
@@ -1 +1 @@
-1.8.0
+1.8.1
diff --git a/cmake/CMakeLists.txt b/cmake/CMakeLists.txt
index 35822e21079d6..0a1187f6b372b 100644
--- a/cmake/CMakeLists.txt
+++ b/cmake/CMakeLists.txt
@@ -1391,8 +1391,11 @@ if (onnxruntime_USE_CUDA)
endif()
enable_language(CUDA)
message( STATUS "CMAKE_CUDA_COMPILER_VERSION: ${CMAKE_CUDA_COMPILER_VERSION}")
-
- set(CMAKE_CUDA_STANDARD 14)
+ if (WIN32)
+ set(CMAKE_CUDA_STANDARD 17)
+ else()
+ set(CMAKE_CUDA_STANDARD 14)
+ endif()
file(TO_CMAKE_PATH ${onnxruntime_CUDNN_HOME} onnxruntime_CUDNN_HOME)
set(ONNXRUNTIME_CUDA_LIBRARIES ${CUDA_LIBRARIES})
diff --git a/docs/Versioning.md b/docs/Versioning.md
index 439d5ab7bc453..42ece36e742ed 100644
--- a/docs/Versioning.md
+++ b/docs/Versioning.md
@@ -26,6 +26,7 @@ For more details on ONNX Release versions, see [this page](https://github.com/on
| ONNX Runtime release version | ONNX release version | ONNX opset version | ONNX ML opset version | Supported ONNX IR version | [Windows ML Availability](https://docs.microsoft.com/en-us/windows/ai/windows-ml/release-notes/)|
|------------------------------|--------------------|--------------------|----------------------|------------------|------------------|
+| 1.8.1 | **1.8** down to 1.2 | 13 | 2 | 7 | Windows AI 1.8+ |
| 1.8.0 | **1.8** down to 1.2 | 13 | 2 | 7 | Windows AI 1.8+ |
| 1.7.0 | **1.8** down to 1.2 | 13 | 2 | 7 | Windows AI 1.7+ |
| 1.6.0 | **1.8** down to 1.2 | 13 | 2 | 7 | Windows AI 1.6+ |
diff --git a/docs/python/README.rst b/docs/python/README.rst
index d1681619df165..157c945356c00 100644
--- a/docs/python/README.rst
+++ b/docs/python/README.rst
@@ -8,6 +8,11 @@ For more information on ONNX Runtime, please see `aka.ms/onnxruntime `_
or the `Github project `_.
"""
-__version__ = "1.8.0"
+__version__ = "1.8.1"
__author__ = "Microsoft"
# we need to do device version validation (for example to check Cuda version for an onnxruntime-training package).
diff --git a/onnxruntime/core/providers/cpu/tensor/upsample.h b/onnxruntime/core/providers/cpu/tensor/upsample.h
index c2aea9374b09a..0b48cd6fed396 100644
--- a/onnxruntime/core/providers/cpu/tensor/upsample.h
+++ b/onnxruntime/core/providers/cpu/tensor/upsample.h
@@ -18,8 +18,8 @@ constexpr const char* UpsampleModeCubic = "cubic";
// is a 4x4 matrix
const size_t CubicModeGridLength = 4;
-using GetNearestPixelFunc = std::function;
-using GetOriginalCoordinateFunc = std::function;
+using GetNearestPixelFunc = int64_t(*)(float, bool);
+using GetOriginalCoordinateFunc = float (*)(float, float, float, float, float, float);
enum UpsampleMode {
NN = 0, // nearest neighbour
diff --git a/onnxruntime/core/providers/cuda/cuda_execution_provider.h b/onnxruntime/core/providers/cuda/cuda_execution_provider.h
index c1c35284ee879..e8848e06f2004 100644
--- a/onnxruntime/core/providers/cuda/cuda_execution_provider.h
+++ b/onnxruntime/core/providers/cuda/cuda_execution_provider.h
@@ -178,7 +178,7 @@ class CUDAExecutionProvider : public IExecutionProvider {
p.reset();
});
}
- std::shared_ptr p{std::make_shared()};
+ std::shared_ptr p = std::make_shared();
};
static const std::shared_ptr& PerThreadContextCache() {
diff --git a/package/rpm/onnxruntime.spec b/package/rpm/onnxruntime.spec
index f91ebb9c7ff24..405e57b3cbf8f 100644
--- a/package/rpm/onnxruntime.spec
+++ b/package/rpm/onnxruntime.spec
@@ -1,5 +1,5 @@
Name: onnxruntime
-Version: 1.8.0
+Version: 1.8.1
Release: 1%{?dist}
Summary: onnxruntime