Summary
In this chapter, we introduced the topic of generative AI and its applications, such as ChatGPT, and gave an overview of the main concepts and components involved, such as cloud computing, NLP, and the transformer model. Since its introduction in 2017, the original transformer model has expanded, leading to explosive growth in models and techniques that extend beyond only NLP-type tasks.
We also briefly traced the development of NLP from RNNs and CNNs to the transformer model and explained how transformers overcome the limitations of the former models by using attention mechanisms and parallel processing. We covered how prompts, or user inputs, are processed by the transformer models to generate responses or completions using various variables and scenarios.
Finally, we provided a brief overview of the LLM landscape and how various transformer architectures can be used for a variety of tasks and different use cases, along with their progression, touching on their expansion...