-
Notifications
You must be signed in to change notification settings - Fork 544
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
TensorRT doesn't accelerate #35
Comments
Could you list the code about how you run the test and measure the runtime? And which model did you use? |
@yinghai The experiments, running on a PC with Ubuntu 16.04 and a GTX 1060 GPU, test LeNet (trained by myself) and Inception-7 (downloaded from mxnet-model-gallery) models. At first, I used your
Unfortunately, it sometimes occured some Cuda error, as described before. Then I followed the TensorRT official documents:
The time cost is manually measured by |
This problem is quite similar to #32. |
What max_batch_size are you specifying? TensorRT performance will be best when batch_size = max_batch_size. |
@benbarsdell Thanks for your reply. I tried with |
We should probably separate the error message from the performance problem. I suggest let's get the error condition sorted out first.
So that's from cuDNN. Exactly which version of cuDNN is this? And while we're at it, which CUDA, which TensorRT versions? |
@cliffwoolley I used cuDNN 7, CUDA 9.0 and TensorRT 4.0. All of them are downloaded from nvidia official websites and installed following their instructions. What's more, I built both MXNet and onnx-tensorrt from sources. |
With apologies, can you say exactly which cuDNN version? There have been around ten different released versions numbered like 7.x.y. |
@cliffwoolley Sorry, I just guess it doesn't matters... I recheck my cuDNN version, typing
and it gets
So it seems that the cuDNN version is 7.0.4, right? |
If you're able to try one of the cuDNN 7.1 or 7.2 builds -- and if that doesn't already fix the problem for you -- then we should be able to use the API logging feature that was added in cuDNN 7.1 to chase down where the problem is happening. Thanks, |
@cliffwoolley It works fine after I upgraded to cuDNN 7.2.1......at least, until now. But the memory cost to run the
Many thanks! |
@benbarsdell Any further advice you can offer here? |
@Faldict -- It's a bit of an aside, but apache/mxnet#11325 was merged to the MXNet master branch today. It uses onnx-tensorrt on your behalf under the hood. I wonder if you have a better experience using that higher-level interface? |
@cliffwoolley I have followed that PR from last month. Now that it is merged, I'll try it. |
any updates? Same issue here, tensorrt does not accelerate onnx(converted from pytorch) models |
try to save .trt and load a again |
Does anyone have a repro for this issue with the latest version of TensorRT (7.2)? |
Closing due to inactivity - if you are still having issues with the latest version of onnx-tensorrt feel free to open a new issue. |
Compared with original models, the time cost using tensorrt engine is two times more. So why doesn't it accelerate the running speed? The figure below shows the MXNet model and TensorRT engine's running time per batch.
Sometimes, it occurs such errors:
It's very weird, and I don't know what happened.
The text was updated successfully, but these errors were encountered: