Home Page and Blog of the Multilingual NLP course @ Sapienza University of Rome
The Transformer architecture. Rationale. Self- and cross-attention; keys, queries, values. Encoder and decoder blocks.
No comments:
Post a Comment