Topic: Deploy Deep Learning models as Microservices in minutes
Powering your application with deep learning is no walk in the park, but is certainly attainable with some tricks and good practice. Serving a deep learning model on a production system demands the model to be stable, reproducible, capable of isolation and to behave as a stand-alone package. One possible solution to this is a containerized microservice.
Ideally, serving deep learning microservices should be quick and efficient, without having to dive deep into the underlying algorithms and their implementation. Too good to be true? Not anymore! Together, we will demystify the process of developing, training, and deploying deep learning models as a web microservice using Model Asset Exchange, an open source framework developed in Python at the IBM Center for Open Source Data and AI Technologies (CODAIT).
We will kick off with an overview of how deep learning models are best published as Docker Images on DockerHub, and are best prepared for deployment in local or cloud environments using Kubernetes or Docker. We highlight the following benefits of such an approach:
* Standardized REST API implementation and application-friendly output format (JSON)
* Abstracting out the complex pre and post-processing portions of the model inputs and outputs.
We will walk you through some super cool applications such as automatic image cropping, age estimation from videos/webcam and Veremin – a video theremin. All these applications and the framework itself are open source and we conclude by inviting contributions and opening the gates for you to be a part of this amazing initiative!
UCSF Mission Bay Conference Center, 1675 Owens St, San Francisco, California, United States