Latte: Latent Attention for Linear Time Transformers
Published in NGSM ICML Workshop, 2024
This paper is about a linear latent variable re-parametrisation of attention and combining it with local standard attention
Recommended citation: Dolga, Rares, Marius Cobzarenco, and David Barber. "Latent Attention for Linear Time Transformers." arXiv preprint arXiv:2402.17512 (2024).
Download Paper