6

What is the reason behind the name "Transformers", for Multi Head Self-Attention-based neural networks from Attention is All You Need?

I have been googling this question for a long time, and nowhere I can find any explanation.

Leevo
  • 305
  • 2
  • 9

1 Answers1

2

The authors of the original paper don't provide an explanation, but I suspect it's a combination of:

brazofuerte
  • 1,031
  • 9
  • 24