HiDream-I1 is a text-to-image model officially open-sourced by HiDream-ai on April 7, 2025. The model has 17B parameters and is released under the MIT license, supporting personal projects, scientific research, and commercial use. It currently performs excellently in multiple benchmark tests.

Model Features

Hybrid Architecture Design A combination of Diffusion Transformer (DiT) and Mixture of Experts (MoE) architecture:

  • Based on Diffusion Transformer (DiT), with dual-stream MMDiT modules processing multimodal information and single-stream DiT modules optimizing global consistency.
  • Dynamic routing mechanism flexibly allocates computing resources, enhancing complex scene processing capabilities and delivering excellent performance in color restoration, edge processing, and other details.

Multimodal Text Encoder Integration Integrates four text encoders:

  • OpenCLIP ViT-bigG, OpenAI CLIP ViT-L (visual semantic alignment)
  • T5-XXL (long text parsing)
  • Llama-3.1-8B-Instruct (instruction understanding) This combination achieves SOTA performance in complex semantic parsing of colors, quantities, spatial relationships, etc., with Chinese prompt support significantly outperforming similar open-source models.

Original Model Versions

HiDream-ai provides three versions of the HiDream-I1 model to meet different needs. Below are the links to the original model repositories:

Model NameDescriptionInference StepsRepository Link
HiDream-I1-FullFull version50๐Ÿค— HiDream-I1-Full
HiDream-I1-DevDistilled dev28๐Ÿค— HiDream-I1-Dev
HiDream-I1-FastDistilled fast16๐Ÿค— HiDream-I1-Fast

About This Workflow Example

In this example, we will use the repackaged version from ComfyOrg. You can find all the model files weโ€™ll use in this example in the HiDream-I1_ComfyUI repository.

Before starting, please update your ComfyUI version to ensure itโ€™s at least after this commit to make sure your ComfyUI has native support for HiDream

HiDream-I1 Workflow

The model requirements for different ComfyUI native HiDream-I1 workflows are basically the same, with only the diffusion models files being different.

If you donโ€™t know which version to choose, please refer to the following suggestions:

  • HiDream-I1-Full can generate the highest quality images
  • HiDream-I1-Dev balances high-quality image generation with speed
  • HiDream-I1-Fast can generate images in just 16 steps, suitable for scenarios requiring real-time iteration

For the dev and fast versions, negative prompts are not needed, so please set the cfg parameter to 1.0 during sampling. We have noted the corresponding parameter settings in the relevant workflows.

The full versions of all three versions require a lot of VRAM - you may need more than 27GB of VRAM to run them smoothly. In the corresponding workflow tutorials, we will use the fp8 version as a demonstration example to ensure that most users can run it smoothly. However, we will still provide download links for different versions of the model in the corresponding examples, and you can choose the appropriate file based on your VRAM situation.

Model Installation

The following model files are common files that we will use. Please click on the corresponding links to download and save them according to the model file save location. We will guide you to download the corresponding diffusion models in the corresponding workflows.

text_encoders๏ผš

VAE

  • ae.safetensors This is Fluxโ€™s VAE model, if you have used Fluxโ€™s workflow before, you may have already downloaded this file.

diffusion models We will guide you to download the corresponding model files in the corresponding workflows.

Model file save location

๐Ÿ“‚ ComfyUI/
โ”œโ”€โ”€ ๐Ÿ“‚ models/
โ”‚   โ”œโ”€โ”€ ๐Ÿ“‚ text_encoders/
โ”‚   โ”‚   โ”œโ”€โ”€โ”€ clip_l_hidream.safetensors
โ”‚   โ”‚   โ”œโ”€โ”€โ”€ clip_g_hidream.safetensors
โ”‚   โ”‚   โ”œโ”€โ”€โ”€ t5xxl_fp8_e4m3fn_scaled.safetensors
โ”‚   โ”‚   โ””โ”€โ”€โ”€ llama_3.1_8b_instruct_fp8_scaled.safetensors
โ”‚   โ””โ”€โ”€ ๐Ÿ“‚ vae/
โ”‚   โ”‚   โ””โ”€โ”€ ae.safetensors
โ”‚   โ””โ”€โ”€ ๐Ÿ“‚ diffusion_models/
โ”‚       โ””โ”€โ”€ ...               # We will guide you to install in the corresponding version workflow       

HiDream-I1 Full Version Workflow

1. Model File Download

Please select the appropriate version based on your hardware. Click the link and download the corresponding model file to save it to the ComfyUI/models/diffusion_models/ folder.

2. Workflow File Download

Please download the image below and drag it into ComfyUI to load the corresponding workflow

3. Complete the Workflow Step by Step

Complete the workflow execution step by step

  1. Make sure the Load Diffusion Model node is using the hidream_i1_full_fp8.safetensors file
  2. Make sure the four corresponding text encoders in QuadrupleCLIPLoader are loaded correctly
    • clip_l_hidream.safetensors
    • clip_g_hidream.safetensors
    • t5xxl_fp8_e4m3fn_scaled.safetensors
    • llama_3.1_8b_instruct_fp8_scaled.safetensors
  3. Make sure the Load VAE node is using the ae.safetensors file
  4. For the full version, you need to set the shift parameter in ModelSamplingSD3 to 3.0
  5. For the Ksampler node, you need to make the following settings
    • Set steps to 50
    • Set cfg to 5.0
    • (Optional) Set sampler to lcm
    • (Optional) Set scheduler to normal
  6. Click the Run button, or use the shortcut Ctrl(cmd) + Enter to execute the image generation

HiDream-I1 Dev Version Workflow

1. Model File Download

Please select the appropriate version based on your hardware, click the link and download the corresponding model file to save to the ComfyUI/models/diffusion_models/ folder.

2. Workflow File Download

Please download the image below and drag it into ComfyUI to load the corresponding workflow

3. Complete the Workflow Step by Step

Complete the workflow execution step by step

  1. Make sure the Load Diffusion Model node is using the hidream_i1_dev_fp8.safetensors file
  2. Make sure the four corresponding text encoders in QuadrupleCLIPLoader are loaded correctly
    • clip_l_hidream.safetensors
    • clip_g_hidream.safetensors
    • t5xxl_fp8_e4m3fn_scaled.safetensors
    • llama_3.1_8b_instruct_fp8_scaled.safetensors
  3. Make sure the Load VAE node is using the ae.safetensors file
  4. For the dev version, you need to set the shift parameter in ModelSamplingSD3 to 6.0
  5. For the Ksampler node, you need to make the following settings
    • Set steps to 28
    • (Important) Set cfg to 1.0
    • (Optional) Set sampler to lcm
    • (Optional) Set scheduler to normal
  6. Click the Run button, or use the shortcut Ctrl(cmd) + Enter to execute the image generation

HiDream-I1 Fast Version Workflow

1. Model File Download

Please select the appropriate version based on your hardware, click the link and download the corresponding model file to save to the ComfyUI/models/diffusion_models/ folder.

2. Workflow File Download

Please download the image below and drag it into ComfyUI to load the corresponding workflow

3. Complete the Workflow Step by Step

Complete the workflow execution step by step

  1. Make sure the Load Diffusion Model node is using the hidream_i1_fast_fp8.safetensors file
  2. Make sure the four corresponding text encoders in QuadrupleCLIPLoader are loaded correctly
    • clip_l_hidream.safetensors
    • clip_g_hidream.safetensors
    • t5xxl_fp8_e4m3fn_scaled.safetensors
    • llama_3.1_8b_instruct_fp8_scaled.safetensors
  3. Make sure the Load VAE node is using the ae.safetensors file
  4. For the fast version, you need to set the shift parameter in ModelSamplingSD3 to 3.0
  5. For the Ksampler node, you need to make the following settings
    • Set steps to 16
    • (Important) Set cfg to 1.0
    • (Optional) Set sampler to lcm
    • (Optional) Set scheduler to normal
  6. Click the Run button, or use the shortcut Ctrl(cmd) + Enter to execute the image generation

GGUF Version Models

You need to use the โ€œUnet Loader (GGUF)โ€ node in City96โ€™s ComfyUI-GGUF to replace the โ€œLoad Diffusion Modelโ€ node.

NF4 Version Models