comfyanonymous is now supporting HiDream natively, so I will focus on improving the HiDream Advanced Node and abandon the dependency of diffusers soon(TM).
In the meanwhile update your comfy and you can test the model with less struggle you might have installing this node ;) You need their official HiDream Model and the four text encoders. Sample Workflow is in the Workflow folder
We have now a bunch of forks for this, and with SanDiegoDude/ComfyUI-HiDream-Sampler actively maintaining, feel free to check out his awsome work, which might solve some issues.
I will have a bit of time over easter to work on the node, but it will be updated at my pace and capabilities.
If anyone would like to contribute, I'd LOVE you to reach out or do PRs, as I am not able to solve all the issues alone (Used to be an illustrator not a dev by trade :D )
What is on my list and where you could support me:
- Edit capabilites: Integrate editing capabilities
- GGUF Support: Integrate Calcuis Model
- Local Checkpoints: Get rid of hugging face download logic
- Bat installer: For the standalone/ windows/ portable users of ComfyUI
- Make installation overall easier: currently many users have problems with the installation process. Meanwhile try to follow this video to get you going.
- Multi Image for Img2img: multi in, multi out
- Better uncensored LLM: LLM which does not OOM, so if you have suggestions
- Manual attention checker: At least for the advance mode, choose between sage, sdpa or flash manually
- Beautifiy codebase: Currently a lot of repetition and thus prone to errors
- Cancel generation: Currently generation can't be cancelled with the stop button
- System Prompt presets: Explore good working system prompts for various use cases
- Clean up UI: Clearer naming, tooltips for inputs and rearrange the fields for faster work
- Explore Lora: While the first Loras for HiDream are being explored it would be good to know how to implement them without big performance losses
- Explore HiResFix Capabilites via Img2img: Dig into HiResFix to have a propper Upscale via native HiDream
- fixed uncensored llm support
- fixed pipelines
- fixed image output
- fixed cache overload
- Added multi image generation (up to 8) (Img2img not yet supported!)
- Fixed Sageattention fetch as first attention method.
- Fixed according to burgstall: Not require auto-gptq anymore! Make sure to git pull and pip install -r requirements.txt!
- Added Image2image functionality
- Flash Attention is no longer needed thanks to power88
- added working uncensored Llama Support (Available via HiDream Sampler Advanced) thanks to sdujack2012 but beware its not quantified so you could get an OOM
A custom ComfyUI node for generating images using the HiDream AI model.
- Supports
full
,dev
, andfast
model types. - Configurable resolution and inference steps.
- Uses 4-bit quantization for lower memory usage.
-
Clone this repository into your
ComfyUI/custom_nodes/
directory:git clone https://github.com/lum3on/comfyui_HiDream-Sampler ComfyUI/custom_nodes/comfyui_HiDream-Sampler
-
Install requirements
pip install -r requirements.txt
or for the portable version:
.\python_embeded\python.exe -m pip install -r .\ComfyUI\custom_nodes\comfyui_HiDream-Sampler\requirements.txt
-
Restart ComfyUI.
Steps to install SageAttention 1:
- Install triton. Windows built wheel, download here:
.\python_embeded\python.exe -s -m pip install (Your downloaded whl package)
linux:
python3 -m pip install triton
- Install sageattention package
.\python_embeded\python.exe -s -m pip install sageattention==1.0.6
PyTorch SDPA is automantically installed when you install PyTorch 2 (ComfyUI Requirement). However, if your torch version is lower than 2. Use this command to update to the latest version.
- linux
python3 -m pip install torch torchvision torchaudio
- windows
.\python_embeded\python.exe -s -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu124
Here's some weight that you need to download (Which will be automantically downloaded when running workflow). Please use huggingface-cli to download.
- Llama Text Encoder
Model | Huggingface repo |
---|---|
4-bit Llama text encoder | hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4 |
Uncensored 4-bit Llama text encoder | John6666/Llama-3.1-8B-Lexi-Uncensored-V2-nf4 |
Original Llama text encoder | nvidia/Llama-3.1-Nemotron-Nano-8B-v1 |
- Quantized Diffusion models (Thanks to
azaneko
for the quantized model!)
Model | Huggingface repo |
---|---|
4-bit HiDream Full | azaneko/HiDream-I1-Full-nf4 |
4-bit HiDream Dev | azaneko/HiDream-I1-Dev-nf4 |
4-bit HiDream Fast | azaneko/HiDream-I1-Fast-nf4 |
- Full weight diffusion model (optional, not recommend unless you have high VRAM)
Model | Huggingface repo |
---|---|
HiDream Full | HiDream-ai/HiDream-I1-Full |
HiDream Dev | HiDream-ai/HiDream-I1-Dev |
HiDream Fast | HiDream-ai/HiDream-I1-Fast |
You can download these weights by this command.
huggingface-cli download (Huggingface repo)
For some region that cannot connect to huggingface. Use this command for mirror.
Windows CMD
set HF_ENDPOINT=https://hf-mirror.com
Windows Powershell
$env:HF_ENDPOINT = "https://hf-mirror.com"
Linux
export HF_ENDPOINT=https://hf-mirror.com
- Add the HiDreamSampler node to your workflow.
- Configure inputs: model_type: Choose full, dev, or fast. prompt: Enter your text prompt (e.g., "A photo of an astronaut riding a horse on the moon"). resolution: Select from available options (e.g., "1024 × 1024 (Square)"). seed: Set a random seed. override_steps and override_cfg: Optionally override default steps and guidance scale.
- Connect the output to a PreviewImage or SaveImage node.
- ComfyUI
- GPU (for model inference)
- Models are cached after the first load to improve performance and use 4-bit quantization models from https://github.com/hykilpikonna/HiDream-I1-nf4.
- Ensure you have sufficient VRAM (e.g., 16GB+ recommended for full mode).
Merged with SanDiegoDude/ComfyUI-HiDream-Sampler who implemented a cleaner version for my originial NF4 / fp8 support.
- Added NF4 (Full/Dev/Fast) download and load support
- Added better memory handling
- Added more informative CLI output for TQDM
- Full/Dev/Fast requires roughly 27GB VRAM
- NF4 requires roughly 15GB VRAM
Build upon the original [HiDream-I1]https://github.com/HiDream-ai/HiDream-I1