Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

why no encoder like detr2 #153

Open
xudh1991 opened this issue Jan 24, 2024 · 0 comments
Open

why no encoder like detr2 #153

xudh1991 opened this issue Jan 24, 2024 · 0 comments

Comments

@xudh1991
Copy link

微信截图_20240123154116
I'm very sorry, although I have seen previous inquiries about this issue, I still want to confirm again because I think this position is somewhat unbelievable
As shown in the figure
Personal understanding: Before the Add operation, it corresponds to the position encoding in DETR, and after the Add operation, it corresponds to the transformer encoder in DETR.
But in your 3Dposition aware Features, there are no complex convolution operations, only an F, which is a flattening operation. That is to say, you completely abandoned the transformer encoder instead of using the 3D Position Encoder to replace this process.
The 3D Position Encoder only changes dimensions and does not involve additional learning. Should this be understood as such??
If it is completely discarded, have you ever tried adding a transformer encoder to improve the accuracy of the model

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant