Deploy deep learning models as a web microservice in minutes
Seamlessly serve state-of-the-art deep learning models as web microservices in minutes and create an application around it without having prior deep learning experience.
About This Session
Powering your application with deep learning is no walk in the park but is certainly attainable with some tricks and good practice. Serving a deep learning model on a production system demands the model to be stable, reproducible, capable of isolation and to behave as a stand-alone package. One possible solution to this is a containerized microservice. Ideally, serving deep learning microservices should be quick and efficient, without having to dive deep into the underlying algorithms and their implementation. Too good to be true? Not anymore! Together, we will demystify the process of developing, training, and deploying deep learning models as a web microservice using Model Asset Exchange, an open source framework developed in Python at the IBM Center for Open Source Data and AI Technologies (CODAIT). We will kick off with an overview of how deep learning models are best published as Docker Images on DockerHub and are best prepared for deployment in local or cloud environments using Kubernetes or Docker. We highlight the following benefits of such an approach:
* Standardized REST API implementation and application-friendly output format (JSON)
* Abstracting out the complex pre and post-processing portions of the model inputs and outputs. We will walk you through some super cool applications. All these are open source and we conclude by opening the gates for you to be a part of this amazing initiative!