lucidrains/h-transformer-1d

H-Transformer for Cross-Attention?

Vbansal21 opened this issue ยท 4 comments

Is it possible to use this architecture for Cross attention?

It's not unfortunately :(

What if the the code was altered to make it support cross attention? Would that give any meaningful result?

@Vbansal21 it's not possible, since there is no notion of locality between the source and target

Okay. Well, then closing this issue.