This New Version of FLUX Runs on 6GB VRAM and is 35% Faster - Flux NF4 Install Guide
HTML-код
- Опубликовано: 26 янв 2025
- Flux NF4 is a highly optimized version of the Flux AI model designed to run efficiently on GPUs with as little as 6GB VRAM. It delivers nearly the same image quality as the original model but with significantly faster performance-up to 35% faster-making it perfect for those with lower-end hardware. With built-in VAE and compatibility with platforms like ComfyUI and Forge, Flux NF4 makes high-resolution image generation accessible and quick, even on less powerful systems.
Links:
NF4 Model: huggingface.co...
Workflow: turboflip.de/f...
Commands: git clone github.com/com...
pip install -r requirements.txt
#stablediffusion #flux1 #chatgpt #FluxNF4
flux is heavily trained on high quality images, so even if you want, its hard to get low quality images such as amateur selfie.
I'd rather wait longer than mass produce lower quality images. I usually lower the resolution till i work on my prompt and other settings and once i get what i want, i increase the resolution and then generate more. Using OG Flux dev on just 8gb vram card, takes 2 mins for 1024*1024 23 steps but images are absolutely worth it. Controlnet is a problem though.
So i tried it and it took slightly less than 50% of time compared to OG flux dev, meaning 2x faster while generating similar image, but on my pc i face issues sometimes where generating takes more time than 6x compared to flux dev and 12x more than nf4, it happens every 3-4 images, so sometimes it works, sometimes it falls on its head. Very inconsistent, it will definitely be slower if i generate 100 images compared to regular flux dev. I know it has to do something with vram being full, but flux dev, a 24gb model works consistently on 8gb vram, nothing seems to compare to it.
@@ImShubhamY nf4 was really slow on my 8gb card
Great review! Really started generating quickly! Thanks to the author!
Hey, I could be wrong, but I think that Flux GGUF is even faster?
Error: All input tensors need to be on the same GPU, but found some tensors to not be on a GPU:
[(torch.Size([393216, 1]), device(type='cpu')), (torch.Size([1, 256]), device(type='cuda', index=0)), (torch.Size([1, 3072]), device(type='cuda', index=0)), (torch.Size([12288]), device(type='cpu')), (torch.Size([16]), device(type='cpu'))]
need nvidia card or upgrade nvidia driver
Am I the only one getting missing node type error when trying to load CheckpointLoaderNF4? This is the 2nd guide I follow and it's literally the same thing that I am doing... and it doesn't work for whatever reason.
you have install the custom node "ComfyUI_bitsandbytes_NF4", and its very wip, you have to set comfyui manger to dev channel and sometimes reduce the security in config file from "normal" to "weak" or "normal-"
8:30 when I paste the link, it says it doesn't recognize it and cant download
make sure you have git installed then paste this: git clone github.com/comfyanonymous/ComfyUI_bitsandbytes_NF4.git
@@elpixel6135 it worked but the next one where u paste pip install -r requirements.txt, it says it doesn't recognize "pip". do I have to install something else?
@@elpixel6135 that worked but the next step where you paste pip install -r requirements.txt, it says "pip is not recognized, do I have to install something else?
I’ve heard about a new technology called GGUF. Is it true? How does it compare to NF4?
better
error: KSampler Expected a cuda device, but got: cpu , but i have gpu with cuda enabled
Dragging the image didn't work to get the workflow, is there another way?
Flux pro is wild!
what about text to image ? is this image to image?
my 2060 6GB takes 30 minutes to create a photo :v
does it support img to img and controlnet
On the step "pip install -r requirements.txt" I get the error "'pip' is not recognized as an internal or external command,
operable program or batch file."
How to solve this?
i got this problem too...any solutions?
@@NewsFixTV No, the story ended with me breaking comfi and reinstalling it
I think that Linux and Python installed is necesary. You can verify that.
I wouldn't call it New Version of FLUX ... New version of Flux would be Flux.1.2 or Flux.2 .... I would call that an "Altered Version" or "New Altered Version" or "New Quantified Version" ....
Original flux-dev and flux-snell works with GTX 1060 6GB VRAM already. Slow, but it does the job...
Yeah, I can't get away from the original, nothing else works right.
Which veraion suitable for me,rtx3050 vram 4gb
I have the same GPU as you. How many seconds it takes for you to generate one 1024x1024 flux-dev and flux-snell picture?
@@mikrodizels flux-dev take long to load - 1,500 seconds, but generate the image is only 120 seconds/iterations. Since flux-dev does most jobs in 6-10 steps it is about 15~20 minutes per image generation. I tend to use 1024X768 in either portrait/landscape.
I load on an old HDD and I believe that is why loading the model is so long (well-relatively).
I haven't use flux-snell recently, will check another time.
I have time, it is for fun, so I don't mind much.
CPU is i5 13600K with 32Gb RAM, all stock standard.
I can do flux on my 980ti 6gb. 1280x768 is like 6 minutes an image for me and I'm shocked that I can even generate that size...
Can I use a lora node to generate consistent character images?
No
Thank you
You didn't say how much VRAM the schnell model was using
12GB, you might be able run it with only 8GB of VRAM but it will take longer to generate
NF4 is deprecated in Comfyui
How are you using FLUX Pro on a local machine? Set up an addict please lol.
FAILED: ComfyUI_bitsandbytes_NF4 [EXPERIMENTAL] And if i load this Workflow, its not loading. It says some errors, too: Invalid workflow against zod schema:
Validation error: Required at "last_link_id"; Required at "nodes"; Required at "links"; Required at "version"; Required at "last_node_id"
same issue - thats why I honesly dislike comfyUI and such - always issues with the scripts and such ...
All input tensors need to be on the same GPU, but found some tensors to not be on a GPU:
[(torch.Size([4718592, 1]), device(type='cpu')), (torch.Size([1, 3072]), device(type='cuda', index=0)), (torch.Size([1, 3072]), device(type='cuda', index=0)), (torch.Size([147456]), device(type='cpu')), (torch.Size([16]), device(type='cpu'))]
did you got the solution ?
I deleted comfyui . It has nearly 1.8k issues in github
do you know how I can add a lora loader into this?
SamplerCustomAdvanced
'ForgeParams4bit' object has no attribute 'quant_storage'
which is the fastest model in the case of a 12 GB RAM video card?
Flux schnell
is flux made from germany because the flux "schnell" (fast)?
Yes, Germany is also where the black forest is.
@@burtpanzer i seen it, "Freiburg im Schwarzwald, Deutschland" is from me 110-120km away 🤣
@@8BitRetroRabbit Okay, I'm in California but it was an easy question to answer...
@@8BitRetroRabbit I guess it wasn't mentioned in this video but the German software company that makes Flux is called Schwarzwald. =D
@@burtpanzer i never heard of it befor, so no good marketing for me :D
if your running flux on 6gig with an NF4 model, you may as well use XL at that point.
I been using turbo Dreamshaper and Atrium myself
one thing i don't understand with flux schnell ( i don't use the others ) is that i never have a realistic result using euler. I changed to LCM i it's far better
Yes schnell doesn't give similar image (realistic) despite same settings
You show memory, not vram
VRAM is more limited
Flux pro obviously has better lighting compare to other versions.
My flux dev is sooo slowly in comparison to SD15 including hiresfix and adetailer
flux is a very resource intensive model so it will naturally be slower than sdxl or sd1.5
@@elpixel6135 at this moment i see no advantage of using flux, i have worse results in comparison to sd15 if it comes to realism. and its way slowler i mean x10 slowler by less resolution..
Im running a version of comfyui-zluda because my GPU is amd MSI armor radeon RX 580 with 8g vram. But I need to use flux.1. Please teach me how to. Any video tutorial for comfyui-zluda flux.1 ? Thanks. 🙏🏻
Does it work on AMD?
you might be able to get it working on AMD GPU's but its not that simple. You might want to stick with online image generators
I can use young on flux pro
No tworking with 5700xt. Model is not even loading in to a memory
load with cpu
1:02 Flux schnell looks better and more realistic
sometimes it does look better then the other versions
age xx works always
Seems not to work on Mac at all
use DiffusionBb and save yourself all that comfyUI A1111 nightmare
hi i'm tess from Digiarty Software. Interested in a collab?
Yor process is already outdated and not longer usable. Shoot the video again and repost.
stay wit sd. flux is sfw
Not if you apply a nsfw lora to it.
@@netneo4038 can it?
PonyXL models now are getting am amazing too
What's the best ver of SDXL and model?, what is the best for Auto1111.. I do not like Comfyui.
FLUXTASTIC, FLUX C4PACITOR, and scg-anatomy already exist to fill that gap, with more LoRAs coming out all the time.
I get this message when updating all: FAILED: ComfyUI_bitsandbytes_NF4 [EXPERIMENTAL]