Model Inference

intermediate
ai & ml
Enhanced Content

Definition

Using a trained AI model to make predictions or decisions on new data. Like applying learned knowledge to solve new problems.

Real-World Example

When you upload a photo to identify a plant, the app uses model inference to predict what species it is based on its trained model.

Cloud Provider Equivalencies

All four clouds provide managed ways to deploy trained models and run inference via real-time (online) endpoints and/or batch jobs. They handle scaling, security, monitoring, and versioning around the inference workload.

AWS
Amazon SageMaker (Real-Time Inference, Asynchronous Inference, Batch Transform)
AZ
Azure Machine Learning (Online Endpoints, Batch Endpoints)
GCP
Vertex AI (Online Prediction, Batch Prediction)
OCI
OCI Data Science (Model Deployment)

Explore More Cloud Computing Terms