HiDream Full nf4 quantized #11337
Replies: 2 comments 5 replies
-
Also GGUF version is available now, if someone know how to use in diffusers. Appreciated if some code can be provided. I am confused as there are 4 text encoder and here only 3. Transformer VAE |
Beta Was this translation helpful? Give feedback.
-
I'll use that method as an option, but problem while integrating is newgenai only has Fast model, and also needed Dev and Full. The other one I was going to use in the beginning is https://huggingface.co/azaneko/HiDream-I1-Full-nf4 and wondering if that'd be compatible with this method. I'd also love to get the GGUF implementation too, if that gets figured out, kinda liked that format. BTW, this is for my app AEIONic.com if you're curious. |
Beta Was this translation helpful? Give feedback.
-
base_repo = "newgenai79/HiDream-I1-Fast-bnb-int4"
text_encoder_2, text_encoder_3 - int4
"hugging-quants/Meta-Llama-3.1-8B-Instruct-GPTQ-INT4"
Meta-Llama-3.1-8B-Instruct-GPTQ - int4
base_repo = "newgenai79/HiDream-I1-Fast-bnb-int4"
transformer - int4
Please NOTE above are pre-quantized model(s) repo.
I couldn't test it as it runs OOM on 16 GB (8+8 GB) even for text encoder
Use the same repo for transformer/vae for inference.
If you don't want to use int4 for text_encoder_2, text_encoder_3, use the original repo by specifying base_repo =
@Skquark
#11231 (comment)
Beta Was this translation helpful? Give feedback.
All reactions