Skip to content

txt2img Guide

You can open the txt2img tab to perform text-to-image inference using the combined functionality of the native region of txt2img and the newly added "Amazon SageMaker Inference" panel in the solution. This allows you to invoke cloud resources for txt2img inference tasks.

txt2img user guide

General Inference Scenario

  1. Navigate to txt2img tab, find Amazon SageMaker Inference panel. Sagemaker Inference面板
  2. Enter the required parameters for inference. Similar to local inference, you can customize the inference parameters of the native txt2img, including model name (stable diffusion checkpoint, extra networks:Lora, Hypernetworks, Textural Inversion and VAE), prompts, negative prompts, sampling parameters, and inference parameters. For VAE model switch, navigate to Settings tab, select Stable Diffusion in the left panel, and then select VAE model in SD VAE (choose VAE model: Automatic = use one with same filename as checkpoint; None = use VAE from checkpoint). Settings 面板

    Notice

    The model files used in the inference should be uploaded to the cloud before generate, which can be referred to the introduction of chapter Cloud Assets Management. The current model list displays options for both local and cloud-based models. For cloud-based inference, it is recommended to select models with the sagemaker keyword as a suffix, indicating that they have been uploaded to the cloud for subsequent inference.

  3. Select Stable Diffusion checkpoint model that will be used for cloud inference through Stable Diffusion Checkpoint Used on Cloud, and the button Generate will change to button Generate on Cloud. Generate button面板

    Notice

    This field is mandatory.

  4. Finish setting all the parameters, and then click Generate on Cloud.

  5. Check inference result. Fresh and select the top option among Inference Job dropdown list, the Output section in the top-right area of the txt2img tab will display the results of the inference once completed, including the generated images, prompts, and inference parameters. Based on this, you can perform subsequent workflows such as clicking Save or Send to img2img.

    Note: The list is sorted in reverse chronological order based on the inference time, with the most recent inference task appearing at the top. Each record is named in the format of inference time -> inference id.

generate results

Continuous Inference Scenario

  1. Following the General Inference Scenario, complete the parameter inputs and click Generate on Cloud to submit the initial inference task.
  2. Wait for the appearance of a new Inference IDin the right-side "Output" section.
  3. Once the new Inference ID appears, you can proceed to click Generate on Cloud again for the next inference task.

generate results

Inference Using Extra Model(s) like Lora

  1. Please follow the native version of WebUI, and upload a copy of the required model (including Textual Inversion, Hypernetworks, Checkpoints or Lora models) to local machine.
  2. Upload corresponding models to the cloud, following Upload Models.
  3. Select the required model, adjust the weights of the model in prompts field, and click Generate on Cloud to inference images.

Inference Job Histories

Inference Job displays the latest 10 inference jobs by default, following naming format Time-Type-Status-UUID. Checking Show All will display all inference jobs the account has. Checking Advanced Inference Job filter and apply filters as need will provide users a customized inference job list.