Skip to content
This repository has been archived by the owner on Nov 17, 2023. It is now read-only.

Commit

Permalink
Imagenet inference to nightly fix (#16599)
Browse files Browse the repository at this point in the history
* split to cd and shell

* comment

* lots of prints

* copy binary at correct location

* remove comments

* add mkl lib

* update docker run build function

* set nvidia docker true to run imagenet inference on GPU

* Revert "set nvidia docker true to run imagenet inference on GPU"

This reverts commit 98f8eef.
As we don't need GPU for compilation.
  • Loading branch information
ChaiBapchya authored and marcoabreu committed Oct 25, 2019
1 parent c0e616f commit c574067
Show file tree
Hide file tree
Showing 2 changed files with 8 additions and 7 deletions.
5 changes: 3 additions & 2 deletions ci/docker/runtime_functions.sh
Original file line number Diff line number Diff line change
Expand Up @@ -1481,8 +1481,9 @@ nightly_test_installation() {
nightly_test_imagenet_inference() {
set -ex
echo $PWD
cp /work/mxnet/build/cpp-package/example/imagenet_inference .
/work/mxnet/cpp-package/example/inference/unit_test_imagenet_inference.sh
cp /work/mxnet/build/cpp-package/example/imagenet_inference /work/mxnet/cpp-package/example/inference/
cd /work/mxnet/cpp-package/example/inference/
./unit_test_imagenet_inference.sh
}

#Runs a simple MNIST training example
Expand Down
10 changes: 5 additions & 5 deletions tests/nightly/JenkinsfileForBinaries
Original file line number Diff line number Diff line change
Expand Up @@ -20,7 +20,7 @@

mx_lib = 'lib/libmxnet.so, lib/libmxnet.a, lib/libtvm_runtime.so, lib/libtvmop.so, 3rdparty/dmlc-core/libdmlc.a, 3rdparty/tvm/nnvm/lib/libnnvm.a'
mx_cmake_lib = 'build/libmxnet.so, build/libmxnet.a, build/3rdparty/tvm/libtvm_runtime.so, build/libtvmop.so, build/3rdparty/dmlc-core/libdmlc.a, build/tests/mxnet_unit_tests, build/3rdparty/openmp/runtime/src/libomp.so'
mx_lib_cpp_example = 'lib/libmxnet.so, lib/libmxnet.a, lib/libtvm_runtime.so, lib/libtvmop.so, 3rdparty/dmlc-core/libdmlc.a, 3rdparty/tvm/nnvm/lib/libnnvm.a, build/cpp-package/example/imagenet_inference'
mx_lib_cpp_example_mkl = 'lib/libmxnet.so, lib/libmxnet.a, lib/libtvm_runtime.so, lib/libtvmop.so, 3rdparty/dmlc-core/libdmlc.a, 3rdparty/tvm/nnvm/lib/libnnvm.a, build/cpp-package/example/imagenet_inference, lib/libmkldnn.so.0, lib/libmklml_intel.so'

node('utility') {
// Loading the utilities requires a node context unfortunately
Expand All @@ -34,10 +34,10 @@ core_logic: {
stage('Build') {
parallel 'GPU: CUDA10.1+cuDNN7': {
node(NODE_LINUX_CPU) {
ws('workspace/build-gpu') {
ws('workspace/build-mkldnn-gpu') {
utils.init_git()
utils.docker_run('ubuntu_build_cuda', 'build_ubuntu_gpu_cuda101_cudnn7', false)
utils.pack_lib('gpu', mx_lib_cpp_example)
utils.docker_run('ubuntu_build_cuda', 'build_ubuntu_gpu_mkldnn', false)
utils.pack_lib('gpu', mx_lib_cpp_example_mkl)
}
}
}/*,
Expand Down Expand Up @@ -73,7 +73,7 @@ core_logic: {
'ImageNet Inference: GPU': {
node(NODE_LINUX_GPU) {
ws('workspace/nt-ImageInferenceTest') {
utils.unpack_and_init('gpu', mx_lib_cpp_example)
utils.unpack_and_init('gpu', mx_lib_cpp_example_mkl)
utils.docker_run('ubuntu_nightly_gpu', 'nightly_test_imagenet_inference', true)
}
}
Expand Down

0 comments on commit c574067

Please sign in to comment.