An Introduction to Pretraining Foundation Models
In this chapter, you’ll be introduced to foundation models, the backbone of many artificial intelligence and machine learning systems today. In particular, we will dive into their creation process, also called pretraining, and understand where it’s competitive to improve the accuracy of your models. We will discuss the core transformer architecture underpinning state-of-the-art models such as Stable Diffusion, BERT, Vision Transformers, OpenChatKit, CLIP, Flan-T5, and more. You will learn about the encoder and decoder frameworks, which work to solve a variety of use cases.
In this chapter, we will cover the following topics:
- The art of pretraining and fine-tuning
- The Transformer model architecture
- State-of-the-art vision and language models
- Encoders and decoders