lucidrains/recurrent-interface-network-pytorch

Lacks a `ff` layer?

CiaoHe opened this issue · 1 comments

Hi Phil!
Compared to Algorithm3 Line 4, I found a little inconsistent,
image

latents = self.latents_attend_to_patches(latents, patches, time = t) + latents

maybe need

latents = self.latents_attend_to_patches(latents, patches, time = t) + latents 
latents = self.latents_cross_attn_ff(latents, time=t) + latents

just curious 😁

@CiaoHe Hi He Cao! Yes indeed all attention should be followed by feedforwards, thank you for catching this!