reference
Transformer-based pre-trained language models are categorized into encoder-only models (e.g., BERT) for understanding and classifying text, decoder-only models (e.g., GPT) for generating coherent text, and encoder-decoder models (e.g., T5) for tasks requiring both comprehension and generation.

Authors

Sources

Referenced by nodes (4)