Skip to content

API Inference Process

How-to inference on Stable Diffusion through API

  • Create Endpoint through CreateEndpoint
  • Upload model file through CreateCheckpoint, Please refer to: API Upload Checkpoint Process
  • Select Async inference or Real-time inference

Async inference

  • Create an inference job through CreateInferenceJob
  • Based on the presigned address api_params_s3_upload_url returned by CreatInferenceJob Upload inference parameters
  • Start an inference job through StartInferenceJob
  • Get an inference job through GetInferenceJob, check the status, and stop the request if successful

Real-time inference

  • Create an inference job through CreateInferenceJob
  • Based on the pre signed address api_params_s3_upload_url returned by CreatInferenceJob Upload inference parameters
  • Starting the inference job through StartInferenceJob, the real-time inference job will get the inference result in this interface

How-to inference on ComfyUI through API

Async inference

  • Create Endpoint through CreateEndpoint
  • Create an inference job through CreateExecute
  • Get an inference job through GetExcute, check the status, and stop the request if successful