reference
The paper 'Theoretical limitations of self-attention in neural sequence models' (Transactions of the Association for Computational Linguistics 8) is cited in the survey 'A Survey on the Theory and Mechanism of Large Language Models' regarding self-attention limitations.
Authors
Sources
- A Survey on the Theory and Mechanism of Large Language Models arxiv.org via serper
Referenced by nodes (1)
- self-attention mechanism concept