====================================================================
Published: 18 December 2021
Tags: machine learning
“Transformers architectures are the hottest thing in supervised and unsupervised learning, achieving SOTA results on natural language processing, vision, audio and multimodal tasks”. In this technically heavy article, Lorenzo Pieri introduces transformers, briefly describes how they were able to simplify previously complex deep learning models, and outlines some of their advantages and disadvantages.