918 B
918 B
How-to inference through API
- Create Endpoint through
CreateEndpoint - Upload model file through
CreateCheckpoint, Please refer to:API Upload Checkpoint Process - Select
Async inferenceorReal-time inference
Async inference
- Create an inference job through
CreateInferenceJob - Based on the presigned address
api_params_s3_upload_urlreturned byCreatInferenceJobUpload inference parameters - Start an inference job through
StartInferenceJob - Obtain an inference job through
GetInferenceJob, check the status, and stop the request if successful
Real-time inference
- Create an inference job through
CreateInferenceJob - Based on the pre signed address
api_params_s3_upload_urlreturned byCreatInferenceJobUpload inference parameters - Starting the inference job through
StartInferenceJob, the real-time inference job will obtain the inference result in this interface