API Inference Process
How-to inference on Stable Diffusion through API
- Create Endpoint through
CreateEndpoint
- Upload model file through
CreateCheckpoint, Please refer to: API Upload Checkpoint Process
- Select
Async inference or Real-time inference
Async inference
- Create an inference job through
CreateInferenceJob
- Based on the presigned address
api_params_s3_upload_url returned by CreatInferenceJob Upload inference parameters
- Start an inference job through
StartInferenceJob
- Get an inference job through
GetInferenceJob, check the status, and stop the request if successful
Real-time inference
- Create an inference job through
CreateInferenceJob
- Based on the pre signed address
api_params_s3_upload_url returned by CreatInferenceJob Upload inference parameters
- Starting the inference job through
StartInferenceJob, the real-time inference job will get the inference result in this interface
How-to inference on ComfyUI through API
Async inference
- Create Endpoint through
CreateEndpoint
- Create an inference job through
CreateExecute
- Get an inference job through
GetExcute, check the status, and stop the request if successful