Chapter 6, Text Generation with OpenAI GPT-2 and GPT-3 Models
- A zero-shot method trains the parameters once. (True/False)
False. No the parameters of the model are first trained through as many episodes as necessary. Zero-shot means that downstream tasks are performed without additional fine-tuning.
- Gradient updates are performed when running zero-shot models. (True/False)
False.
- GPT models only have a decoder stack. (True/False)
True.
- It is impossible to train a 117M GPT model on a local machine. (True/False)
False. We trained one in this chapter.
- It is impossible to train the GPT-2 model with a specific dataset. (True/False)
False. We trained one in this chapter.
- A GPT-2 model cannot be conditioned to generate text. (True/False)
False. We implemented this in this chapter.
- A GPT-2 model can analyze the context of input and produce completion content. (True/False)
True.
- We cannot interact with...