-
Notifications
You must be signed in to change notification settings - Fork 96
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Fuse quantizelinear for skip layers using multioutput fusions #3791
Labels
Comments
Here is some example code used for the experiment.
|
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Follow up from PR #3782
Ex resnet quantized graph after above PR:
Doing some experimental work, it turns out that we get a slight perf boost from moving the skip-connection quantize op before the step op, and fusing it into the previous conv-pointwise kernel. This should probably be done as 2 steps:
mlir_quant_convolution_dequantizelinear_dequantizelinear_add_add_relu_quantizelinear_quantizelinear
These should be done in this order since swapping the order of quantize op and step op is generally not preferable unless doing it for this fusion.
The text was updated successfully, but these errors were encountered: