Requests to Models

Requests can be made using the user interface via different options.

Make a prediction

The prediction can be made from the UI directly by pasting or uploading content. The response is then shown on the screen.


Here the path to model is displayed and an option is provided to export a curl command for manual requests.

SeldonDeployment models default to the Seldon protocol and URL form. Alternatively, the tensorflow protocol can be used and then Deploy will infer a tensorflow URL. With Seldon the model name needs to be specified as a parameter in the manifest. Content of tensorflow requests is different as explained in a seldon core notebook.

Load Test

This initiates a loadtest, which in the background is implemented using hey and exposes the same options as that tool


The load test runs inside the cluster so can take time to be provisioned.