-
Notifications
You must be signed in to change notification settings - Fork 643
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[WIP] Support deploy MMRazor quantized model #1471
Open
pppppM
wants to merge
23
commits into
open-mmlab:dev-1.x
Choose a base branch
from
pppppM:adapt_razor_quantize
base: dev-1.x
Could not load branches
Branch not found: {{ refName }}
Loading
Could not load tags
Nothing to show
Loading
Are you sure you want to change the base?
Some commits from the old base branch may be removed from the timeline,
and old review comments may become outdated.
Open
Conversation
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
grimoire seems not to be a GitHub user. You need a GitHub account to be able to sign the CLA. If you have already a GitHub account, please add the email address used for this commit to your account. You have signed the CLA already but the status is still pending? Let us recheck it. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Motivation
The related pr in
MMRazor
is open-mmlab/mmrazor#365MMRazor is developing quantization algorithms, including PTQ and QAT.
This PR is a draft code to deploy
MMRazor
quantization model inMMDeploy
, mainly with the following two points.Export FX Graph
MMRazor quantized model is FX graph, and the current function rewriter cannot handle FX graph correctly.
The function rewriter has been fine-tuned in this PR, which can handle FX graph correctly.
Export Quantized ONNX
Different backends have different ONNX formats for quantized models, TensorRT and Openvion's quantized onnx exporters are implemented in this pr.
Modification
Function Rewriter
The original function rewriter is a wrapper, and the first arg is
ctx
.In order to process FX Graph, wrapper is no longer used in this pr.
The original function is directly replaced by rewritten function, and
ctx
is removed from args.ctx
becomes a global variable.Quantize ONNX Exporter
This pr adds a fake quant symbolic op, with which a temporary non-running onnx can be exported.
Then, different backends quantize onnx exporter will convert it to final deployed onnx.