claim
Large Language Models (LLMs) are based on the transformer architecture, which excels in handling long sequences due to its self-attention mechanism.

Authors

Sources

Referenced by nodes (3)