Transformer Recipe
Transformers have accelerated the development of new techniques and models for natural language processing (NLP) tasks. While it has mostly been used for NLP tasks, it is now seeing heavy adoption in other areas such as computer vision and reinforcement learning. That makes it one of the most important modern concepts to understand and be able to apply.
I am aware that a lot of machine learning and NLP students and practitioners are keen on learning about transformers. Therefore, I have prepared a study guide in the form of a list of resources and study materials to help guide students interested in learning about the world of Transformers.
To begin with, I have prepared a few links to materials that I used to better understand and implement transformer models from scratch.
High-level Introduction
First, try to get a very high-level introduction about transformers. Some references worth looking at:
๐ Introduction to Transformer - Lecture Notes (Elvis Saravia)
๐ Transformers From Scratch (Brandon Rohrer)
๐ How Transformers work in deep learning and NLP: an intuitive introduction (AI Summer)
๐ Stanford CS25 - Transformers United
๐ Deep Learning for Language Understanding (DeepMind)
๐ Transformer models: an introduction and catalog (Xavier Amatriain)
The Transformer Explained
Jay Alammar's illustrated explanations are exceptional. Once you get that high-level understanding of transformers, you can jump into this popular detailed and illustrated explanation of transformers:
๐ The Illustrated Transformer
This next article also breaks down Transformers into its components, explaining and illustrating in detail what each part does:
๐ Breaking Down the Transformer
Technical Summary
At this point, you may be looking for a technical summary and overview of transformers. Lilian Weng's blog posts are a gem and provide concise technical explanations/summaries:
๐ The Transformer Family Version 2.0
Implementation
After the theory, it's important to test the knowledge. I typically prefer to understand things in more detail so I prefer to implement algorithms from scratch. For implementing transformers, I mainly relied on this tutorial:
๐ The Annotated Transformer | (Google Colab | GitHub)
๐ Language Modeling with nn.Transformer and TorchText
If you are looking for in-depth implementations on some of the latest transformers, you might also find the Papers with Code methods collection for Transformers useful.
Attention Is All You Need
This paper by Vaswani et al. introduced the Transformer architecture. Read it after you have a high-level understanding and want to get into the details. Pay attention to other references in the paper for diving deep.
๐ Attention Is All You Need
Applying Transformers
After some time studying and understanding the theory behind transformers, you may be interested in applying them to different NLP projects or research. At this time, your best bet is the Transformers library by HuggingFace.
๐ Transformers
The Hugging Face Team has also published a new book on NLP with Transformers, so you might want to check that out here.
Reading List on LLMs
As a bonus, here is a great reading list on LLMs by Sebastian Raschka.
๐ Understanding Large Language Models -- A Transformative Reading List
Feel free to suggest study material. In the next update, I am looking to add a more comprehensive collection of Transformer applications and papers. In addition, a code implementation for easy experimentation is coming as well. Stay tuned!
To get regular updates on new ML and NLP resources, follow me on Twitter.