Home
About
Light
Dark
Automatic
English
中文 (简体)
Post
LLaMA Code Analysis
Learn how modern transformers work by analyzing the official implementation of LLaMA
T5 Paper Notes
T5 is an encoder-decoder transformer. It utilizes multi-task learning in the training process. By studying the T5 paper, we can learn many interesting details about pre-training of transformer models.