Skip to content

Files

Latest commit

 

History

History
21 lines (11 loc) · 1019 Bytes

File metadata and controls

21 lines (11 loc) · 1019 Bytes

Model Serving Guide

Deployment

Assuming you have already created a model in the Model Registry, a deployment can now be created to prepare a model artifact for this model and make it accessible for running predictions behind a REST endpoint. Follow the Deployment Creation Guide to create a Deployment for your model.

Predictor

Predictors are responsible for running a model server that loads a trained model, handles inference requests and returns predictions, see the Predictor Guide.

Transformer

Transformers are used to apply transformations on the model inputs before sending them to the predictor for making predictions using the model, see the Transformer Guide.

Inference Batcher

Configure the predictor to batch inference requests, see the Inference Batcher Guide.

Inference Logger

Configure the predictor to log inference requests and predictions, see the Inference Logger Guide.