claim
Wu et al. (2025a) proposed the Parallel Loop Transformer (PLT) architecture, which is designed to improve computational efficiency when leveraging recurrence in language models.
Authors
Sources
- A Survey on the Theory and Mechanism of Large Language Models arxiv.org via serper
Referenced by nodes (1)
- Language Model concept