From the course: Cloud-Based AI Solution Design Patterns

Unlock this course with a free trial

Join today to access over 24,900 courses taught by industry experts.

Containerized AI-model deployment

Containerized AI-model deployment

- AI models are typically viewed as parts of an overall AI system containing the logic that gets executed by the system's inference engine. But as with any distributed application, it sometimes makes sense to look at how different components of the application may be able to benefit from being separated so that they can be scaled and managed independently. This is especially the case with a critical component such as the AI model. The containerized AI model deployment pattern essentially provides us with the option of putting a model together with its dependencies into a container. Using an appropriate underlying platform such as an orchestration platform, we can then create instances of the same containerized model. Why would we want to do this? Well, in addition to being able to scale the model dynamically in cloud environments, it also helps guarantee its availability and further supports efficient resource allocation as instances of the containerized model can be created and…

Contents