claim
Looped architectures in large language models can simulate Chain-of-Thought (CoT) internally through 'latent thoughts', which can efficiently substitute for explicit token generation.
Authors
Sources
- A Survey on the Theory and Mechanism of Large Language Models arxiv.org via serper
Referenced by nodes (2)
- Large Language Models concept
- chain-of-thought concept