wzzheng/TPVFormer

Question about Cross-view Hybrid attention

jianingwangind opened this issue · 3 comments

Thanks for sharing the great work.

Regarding to Cross-view Hybrid attention, is it only apllied for the HW top plane?

The query is itself, key and value are both none while later in cross-view hybrid attention the value is set to be the concatenation of queries

value = torch.cat([query, query], 0)

I have the same question, it looks like there is no interaction between the features of the three planes.

Thanks for your interest in our work.
Your understanding of the code is correct. That is, in TPVFormer04, cross-view hybrid attention is enabled only in the HW plane, thus degrading to self-attention.

@huang-yh Thanks for your reply. May i further ask the idea behind this? Similar performance when disabling the attention in the other two planes?