Skip to content

Files

Latest commit

1d50363 · Jun 11, 2020

History

History

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020
Jun 11, 2020

Chapter 9 - Scalable Inference Serving on Cloud with TensorFlow Serving and KubeFlow

Note: All images in this directory, unless specified otherwise, are licensed under CC BY-NC 4.0.

Figure List

Figure number Description Notes
9-1 A high-level overview and comparison of different inference serving options
9-2 Navigate to http://localhost:5000/hello within a web browser to view the “Hello World!” web page
9-3 Listing page for machine learning models on the Google Cloud ML Engine dashboard
9-4 Model creation page on Google Cloud ML Engine
9-5 Model listings page on Google Cloud ML Engine
9-6 Details page of the just-created Dog/Cat classifier
9-7 Creating a new version for a machine learning model
9-8 Creating a new Google Cloud Storage bucket within the ML model version creation page
9-9 Google Cloud Storage Browser page showing the uploaded Dog/Cat classifier model in TensorFlow format
9-10 Add the URI for the model you uploaded to Google Cloud Storage
9-11 An end-to-end pipeline illustrated in KubeFlow
9-12 Creating a new Jupyter Notebook server on KubeFlow
9-13 Creating a KubeFlow deployment on GCP using the browser
9-14 Google Cloud ML Engine showing incoming queries and latency of serving the calls, with end-to-end latency at user’s end of about 3.5 seconds
9-15 Cost comparison of infrastructure as a service (Google Cloud ML Engine) versus building your own stack over virtual machines (Azure VM) (costs as of August 2019)