reference
Encoder-decoder architectures, such as T5 or BART (Bidirectional and Auto-Regressive Transformers), use an encoder to create a context-rich representation of the input sequence, which the decoder then uses to generate an output sequence, making them flexible for tasks like translation, summarization, and question answering.

Authors

Sources

Referenced by nodes (4)