This part introduces ways to improve the efficiency and speed of a model and its pipeline. We will start with the concept of model runtime, and then model optimization, followed by using TensorFlow Serving to serve models as a Docker container via a RESTful API.
This section comprises the following chapters:
18.119.131.178