Fix the QuantizedAVX2 build issue#26854
Conversation
|
Any perf checks? |
facebook-github-bot
left a comment
There was a problem hiding this comment.
@llyfacebook has imported this pull request. If you are a Facebook employee, you can view this diff on Phabricator.
Yes, I found it is even faster than the original one. (For bilinear2d interpolate case.) |
|
I got some undeclared identifier errors when build locally: |
|
@llyfacebook yeah, QuantizeAVX2 operates on vectors of 32, but you were feeding it vectors of 8, so it was just running scalar code :p |
|
@llyfacebook I'd expect another 2x speedup if you switch to doing the float operations 4-wide and using QuantizeAVX2 again |
hx89
left a comment
There was a problem hiding this comment.
LGTM! My local build pass after rebasing.
facebook-github-bot
left a comment
There was a problem hiding this comment.
@llyfacebook is landing this pull request. If you are a Facebook employee, you can view this diff on Phabricator.
|
@llyfacebook merged this pull request in 428204d. |
Summary: The QuantizedAVx2 does not support the int32 type. We switch to use at::quantize_vec function instead. Pull Request resolved: pytorch/pytorch#26854 Differential Revision: D17609872 Pulled By: llyfacebook fbshipit-source-id: b4a77d93ce0ebfef696506b5cdbe3e91fe44bb36
Summary: The QuantizedAVx2 does not support the int32 type. We switch to use at::quantize_vec function instead. Pull Request resolved: pytorch#26854 Differential Revision: D17609872 Pulled By: llyfacebook fbshipit-source-id: b4a77d93ce0ebfef696506b5cdbe3e91fe44bb36
Summary: The QuantizedAVx2 does not support the int32 type. We switch to use at::quantize_vec function instead. Pull Request resolved: pytorch#26854 Differential Revision: D17609872 Pulled By: llyfacebook fbshipit-source-id: b4a77d93ce0ebfef696506b5cdbe3e91fe44bb36
The QuantizedAVx2 does not support the int32 type. We switch to use at::quantize_vec function instead.