TensorFlow Serving is a flexible, high-performance serving system for machine learning models, designed for production environments. TensorFlow Serving makes it easy to deploy new algorithms and experiments, while keeping the same server architecture and APIs. Learn more about Tensorflow on their site: ****
Quick Start Guide
Running TensorFlow Serving to serve the TensorFlow ResNet model is, as usual, a single line trick.
For a full explanation of how to use Tensorflow Serving please take a look at the documentation site:
Step 1 - Prepare your client environment
To run any experiment against a remote inference server you have to have tensorflow-serving-api installed on your machine. As per official documentation here:
As reference
pip install tensorflow-serving-api
Step 2
Clone the TensorFlow repository where we will find the test scripts
git clone https://github.com/tensorflow/serving
cd serving
Step 3
Find your cluster IP where Tensorflow Serving service is exposed
kubectl describe service tf-server-service -n tf-serving