Transformers

12345678910
Across
  1. 3. Recurrent Neural Networks
  2. 5. Encoder output
  3. 9. Decoder mechanism
  4. 10. Processes input sequence
Down
  1. 1. Mechanism for long-range dependencies
  2. 2. Generates output sequence
  3. 4. Natural Language Processing
  4. 6. Attention-based model
  5. 7. Transformer excels at this
  6. 8. RNN weakness