Introduction to Inference

After an AI model is developed, you can use it to create an AI application and quickly deploy the application as an inference service. The AI inference capabilities can be integrated into your IT platform by calling APIs.

Figure 1 Inference

Deploying an AI Application as a Service

After an AI application is created, you can deploy it as a service on the Deploy page. ModelArts supports the following deployment types:
  • Real-time service

    Deploy an AI application as a web service with real-time test UI and monitoring supported.

  • Batch service

    Deploy an AI application as a batch service that performs inference on batch data and automatically stops after data processing is complete.