lucidrains/x-transformers

"Stabilizing Transformer Training by Preventing Attention Entropy Collapse" improvement to ViT

catid opened this issue · 1 comments

From Apple paper: "Stabilizing Transformer Training by Preventing Attention Entropy Collapse" https://github.com/apple/ml-sigma-reparam

Here's my experiment:

catid/cifar10deepspeed@72a0b78

It works! Does a bit better than the default model: 85.6% vs 85.93% and gets rid of those awkward norms.

I also tried changing other Linear layers in SPT and in the output projection but that breaks the model. So these changes only seem to make sense inside the transformer.

On Twitter the authors said you can apply it to other layers like this: https://github.com/catid/cifar10deepspeed/pull/1/files
But it doesn't work in my script for whatever reason and seems like a bit of a hack.

@catid hey! thanks for sharing! probably a bit too complicated to introduce to this repository, but a good trick to know