Summary
In this chapter, the focus was on deploying your model as a REST endpoint to support real-time inferencing use cases. We saw that we are able to leverage AMLS Studio for a low-code deployment experience. We also leveraged SDK v2 to deploy models to managed online endpoints. We continued by deploying models through CLI v2 to support model deployment for real-time inferencing. These sections demonstrated deploying real-time web services through low-code, code-first, and configuration-driven approaches. These capabilities empower you to deploy in a variety of ways.
In the next chapter, we will learn how to leverage batch-inferencing to support our use cases.