-
Notifications
You must be signed in to change notification settings - Fork 1.8k
[Model Compression] fix compression doc and update a notebook example #3955
Conversation
@@ -31,7 +31,7 @@ The specification of configuration can be found `here <./Tutorial.rst#specify-th | |||
Step2. Choose a pruner and compress the model |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I suggest that we can put the notebook into documentation to increase visibility.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
good suggestion, add this into the doc.
} | ||
], | ||
"source": [ | ||
"from nni.algorithms.compression.pytorch.quantization import DoReFaQuantizer\n", |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Recommend using QAT_Quantizer
instead of DoReFaQuantizer
since quantization speedup only support QAT_Quantizer
right now.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
got it, change to QAT_Quantizer
No description provided.