Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why does the offset_normalizer use switched WH order in deformable attention #3200

Open
Beanocean opened this issue Nov 5, 2024 · 0 comments

Comments

@Beanocean
Copy link

In the deformable attention, the spatial shapes represented by height (H) and width (W). Why is it necessary to exchange the offset_normalizer to the witdh(W) height(H) order here?

offset_normalizer = torch.stack(
[spatial_shapes[..., 1], spatial_shapes[..., 0]], -1)

Originally posted by @Beanocean in #3197

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant