Testing Inferenceยค
You can use the example script here and the make command make test-tf-serving to test the inference of a model served by TensorFlow Serving.
The make command spins up a tensorflow serving Docker container and the script sends a request to the server to get a prediction.
- Depending on your model you will need to edit the inputs to the REST call in the script.
- If you are on an M1/M2 chip then you need to ensure you have turned off
Rosettain Docker Desktop settings as seen below:
An E2E example can be run by using the make test-end-to-end command, hitting enter where necessary to accept the defaults.
