Investigating the potential of downstream tasks
Transformers, like humans, can be fine-tuned to perform downstream tasks by inheriting the properties of a pretrained model. The pretrained model provides its architecture and language representations through its parameters.
A pretrained model trains on key tasks to acquire a general knowledge of the language. A fine-tuned model trains on downstream tasks. Not every transformer model uses the same tasks for pretraining. But, potentially, all tasks can be pretrained or fine-tuned.
Organizing downstream tasks provides a scientific framework for implementing and measuring NLP. However, every NLP model needs to be evaluated with a standard method.
This section will first go through some of the key measurement methods. Then, we will go through some of the main benchmark tasks and datasets.
Let’s start by going through some of the key metric methods.
Evaluating models with metrics
It is impossible to compare...