Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

use tensorrt to inference #4

Open
Yutong-gannis opened this issue Mar 18, 2023 · 1 comment
Open

use tensorrt to inference #4

Yutong-gannis opened this issue Mar 18, 2023 · 1 comment

Comments

@Yutong-gannis
Copy link

how to use tensorrt to do inference

@qinjian623
Copy link
Contributor

To use TensorRT for inference with your PyTorch model, please follow these steps:

  1. Convert your PyTorch model into an ONNX format file. If you haven't done this before, please refer to the PyTorch official documentation: https://pytorch.org/tutorials/advanced/super_resolution_with_onnxruntime.html.
  2. Set up your TensorRT environment. The easiest way to do this is to use a Docker container from NVIDIA: https://ngc.nvidia.com/catalog/containers/nvidia:tensorrt.
  3. Once your TensorRT environment is set up, you can use the trtexec command to perform a random input inference test and obtain a detailed latency report. Follow the instructions provided by the trtexec command to perform the test.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants