Introducing the CLIP model
We have explored computer vision in Chapter 11, Categorizing Images of Clothing with Convolutional Neural Networks, and NLP in Chapter 12, Making Predictions with Sequences Using Recurrent Neural Networks, and Chapter 13, Advancing Language Understanding and Generation with the Transformer Models. In this chapter, we will delve into a model that bridges the realms of computer vision and NLP, the Contrastive Language–Image Pre-Training (CLIP) model developed by OpenAI. Unlike traditional models that are specialized for either computer vision or natural language processing, CLIP is trained to understand both modalities (image and text) in a unified manner. Hence, CLIP excels at understanding and generating relationships between images and natural language.
A modality in ML/AI is a specific way of representing information. Common modalities include text, images, audio, video, and even sensor data.
Excited to delve into the workings...