Editing Help:Editing
Jump to navigation
Jump to search
The edit can be undone. Please check the comparison below to verify that this is what you want to do, and then publish the changes below to finish undoing the edit.
Latest revision | Your text | ||
Line 152: | Line 152: | ||
<references /></pre> | <references /></pre> | ||
{{Paper|title=Attention Is All You Need|authors=Ashish Vaswani et al.|url=https://arxiv.org/abs/1706.03762|tldr=They propose a new network architecture called the Transformer, based solely on attention mechanisms, which outperforms existing models in machine translation tasks while being more parallelizable and requiring less training time.|publication=arXiv|year=2017}} | {{Paper|title=Attention Is All You Need|authors=Ashish Vaswani et al.|url=https://arxiv.org/abs/1706.03762|tldr=They propose a new network architecture called the Transformer, based solely on attention mechanisms, which outperforms existing models in machine translation tasks while being more parallelizable and requiring less training time.|publication=arXiv|year=2017}} | ||
==== References | ==== References === | ||
<references /> | <references /> | ||