509 Commits (c2bb34d865a4c512091e6036d421dd26d0889b25)

Author SHA1 Message Date
comfyanonymous 326577d04c Allow cancelling of everything with a progress bar. 1 year ago
comfyanonymous f88f7f413a Add a ConditioningSetAreaPercentage node. 1 year ago
comfyanonymous 1938f5c5fe Add a force argument to soft_empty_cache to force a cache empty. 1 year ago
comfyanonymous 7746bdf7b0 Merge branch 'generalize_fixes' of https://github.com/simonlui/ComfyUI 1 year ago
Simon Lui 2da73b7073 Revert changes in comfy/ldm/modules/diffusionmodules/util.py, which is unused. 2 years ago
comfyanonymous a74c5dbf37 Move some functions to utils.py 2 years ago
Simon Lui 4a0c4ce4ef Some fixes to generalize CUDA specific functionality to Intel or other GPUs. 2 years ago
comfyanonymous 77a176f9e0 Use common function to reshape batch to. 2 years ago
comfyanonymous 7931ff0fd9 Support SDXL inpaint models. 2 years ago
comfyanonymous 0e3b641172 Remove xformers related print. 2 years ago
comfyanonymous 5c363a9d86 Fix controlnet bug. 2 years ago
comfyanonymous cfe1c54de8 Fix controlnet issue. 2 years ago
comfyanonymous 1c012d69af It doesn't make sense for c_crossattn and c_concat to be lists. 2 years ago
comfyanonymous 7e941f9f24 Clean up DiffusersLoader node. 2 years ago
Simon Lui 18617967e5
Fix error message in model_patcher.py
Found while tinkering.
2 years ago
comfyanonymous fe4c07400c Fix "Load Checkpoint with config" node. 2 years ago
comfyanonymous f2f5e5dcbb Support SDXL t2i adapters with 3 channel input. 2 years ago
comfyanonymous 15adc3699f Move beta_schedule to model_config and allow disabling unet creation. 2 years ago
comfyanonymous bed116a1f9 Remove optimization that caused border. 2 years ago
comfyanonymous 65cae62c71 No need to check filename extensions to detect shuffle controlnet. 2 years ago
comfyanonymous 4e89b2c25a Put clip vision outputs on the CPU. 2 years ago
comfyanonymous a094b45c93 Load clipvision model to GPU for faster performance. 2 years ago
comfyanonymous 1300a1bb4c Text encoder should initially load on the offload_device not the regular. 2 years ago
comfyanonymous f92074b84f Move ModelPatcher to model_patcher.py 2 years ago
comfyanonymous 4798cf5a62 Implement loras with norm keys. 2 years ago
comfyanonymous b8c7c770d3 Enable bf16-vae by default on ampere and up. 2 years ago
comfyanonymous 1c794a2161 Fallback to slice attention if xformers doesn't support the operation. 2 years ago
comfyanonymous d935ba50c4 Make --bf16-vae work on torch 2.0 2 years ago
comfyanonymous a57b0c797b Fix lowvram model merging. 2 years ago
comfyanonymous f72780a7e3 The new smart memory management makes this unnecessary. 2 years ago
comfyanonymous c77f02e1c6 Move controlnet code to comfy/controlnet.py 2 years ago
comfyanonymous 15a7716fa6 Move lora code to comfy/lora.py 2 years ago
comfyanonymous ec96f6d03a Move text_projection to base clip model. 2 years ago
comfyanonymous 30eb92c3cb Code cleanups. 2 years ago
comfyanonymous 51dde87e97 Try to free enough vram for control lora inference. 2 years ago
comfyanonymous e3d0a9a490 Fix potential issue with text projection matrix multiplication. 2 years ago
comfyanonymous cc44ade79e Always shift text encoder to GPU when the device supports fp16. 2 years ago
comfyanonymous a6ef08a46a Even with forced fp16 the cpu device should never use it. 2 years ago
comfyanonymous 00c0b2c507 Initialize text encoder to target dtype. 2 years ago
comfyanonymous f081017c1a Save memory by storing text encoder weights in fp16 in most situations.
Do inference in fp32 to make sure quality stays the exact same.
2 years ago
comfyanonymous afcb9cb1df All resolutions now work with t2i adapter for SDXL. 2 years ago
comfyanonymous 85fde89d7f T2I adapter SDXL. 2 years ago
comfyanonymous cf5ae46928 Controlnet/t2iadapter cleanup. 2 years ago
comfyanonymous 763b0cf024 Fix control lora not working in fp32. 2 years ago
comfyanonymous 199d73364a Fix ControlLora on lowvram. 2 years ago
comfyanonymous d08e53de2e Remove autocast from controlnet code. 2 years ago
comfyanonymous 0d7b0a4dc7 Small cleanups. 2 years ago
Simon Lui 9225465975 Further tuning and fix mem_free_total. 2 years ago
Simon Lui 2c096e4260 Add ipex optimize and other enhancements for Intel GPUs based on recent memory changes. 2 years ago
comfyanonymous e9469e732d --disable-smart-memory now disables loading model directly to vram. 2 years ago
comfyanonymous c9b562aed1 Free more memory before VAE encode/decode. 2 years ago
comfyanonymous b80c3276dc Fix issue with gligen. 2 years ago
comfyanonymous d6e4b342e6 Support for Control Loras.
Control loras are controlnets where some of the weights are stored in
"lora" format: an up and a down low rank matrice that when multiplied
together and added to the unet weight give the controlnet weight.

This allows a much smaller memory footprint depending on the rank of the
matrices.

These controlnets are used just like regular ones.
2 years ago
comfyanonymous 39ac856a33 ReVision support: unclip nodes can now be used with SDXL. 2 years ago
comfyanonymous 76d53c4622 Add support for clip g vision model to CLIPVisionLoader. 2 years ago
Alexopus e59fe0537a
Fix referenced before assignment
For https://github.com/BlenderNeko/ComfyUI_TiledKSampler/issues/13
2 years ago
comfyanonymous be9c5e25bc Fix issue with not freeing enough memory when sampling. 2 years ago
comfyanonymous ac0758a1a4 Fix bug with lowvram and controlnet advanced node. 2 years ago
comfyanonymous c28db1f315 Fix potential issues with patching models when saving checkpoints. 2 years ago
comfyanonymous 3aee33b54e Add --disable-smart-memory for those that want the old behaviour. 2 years ago
comfyanonymous 2be2742711 Fix issue with regular torch version. 2 years ago
comfyanonymous 89a0767abf Smarter memory management.
Try to keep models on the vram when possible.

Better lowvram mode for controlnets.
2 years ago
comfyanonymous 2c97c30256 Support small diffusers controlnet so both types are now supported. 2 years ago
comfyanonymous 53f326a3d8 Support diffusers mini controlnets. 2 years ago
comfyanonymous 58f0c616ed Fix clip vision issue with old transformers versions. 2 years ago
comfyanonymous ae270f79bc Fix potential issue with batch size and clip vision. 2 years ago
comfyanonymous a2ce9655ca Refactor unclip code. 2 years ago
comfyanonymous 9cc12c833d CLIPVisionEncode can now encode multiple images. 2 years ago
comfyanonymous 0cb6dac943 Remove 3m from PR #1213 because of some small issues. 2 years ago
comfyanonymous e244b2df83 Add sgm_uniform scheduler that acts like the default one in sgm. 2 years ago
comfyanonymous 58c7da3665 Gpu variant of dpmpp_3m_sde. Note: use 3m with exponential or karras. 2 years ago
comfyanonymous ba319a34e4 Merge branch 'dpmpp3m' of https://github.com/FizzleDorf/ComfyUI 2 years ago
FizzleDorf 3cfad03a68 dpmpp 3m + dpmpp 3m sde added 2 years ago
comfyanonymous 585a062910 Print unet config when model isn't detected. 2 years ago
comfyanonymous c8a23ce9e8 Support for yet another lora type based on diffusers. 2 years ago
comfyanonymous 2bc12d3d22 Add --temp-directory argument to set temp directory. 2 years ago
comfyanonymous c20583286f Support diffuser text encoder loras. 2 years ago
comfyanonymous cf10c5592c Disable calculating uncond when CFG is 1.0 2 years ago
comfyanonymous 1f0f4cc0bd Add argument to disable auto launching the browser. 2 years ago
comfyanonymous d8e58f0a7e Detect hint_channels from controlnet. 2 years ago
comfyanonymous c5d7593ccf Support loras in diffusers format. 2 years ago
comfyanonymous 1ce0d8ad68 Add CMP 30HX card to the nvidia_16_series list. 2 years ago
comfyanonymous c99d8002f8 Make sure the pooled output stays at the EOS token with added embeddings. 2 years ago
comfyanonymous 4a77fcd6ab Only shift text encoder to vram when CPU cores are under 8. 2 years ago
comfyanonymous 3cd31d0e24 Lower CPU thread check for running the text encoder on the CPU vs GPU. 2 years ago
comfyanonymous 2b13939044 Remove some useless code. 2 years ago
comfyanonymous 95d796fc85 Faster VAE loading. 2 years ago
comfyanonymous 4b957a0010 Initialize the unet directly on the target device. 2 years ago
comfyanonymous c910b4a01c Remove unused code and torchdiffeq dependency. 2 years ago
comfyanonymous 1141029a4a Add --disable-metadata argument to disable saving metadata in files. 2 years ago
comfyanonymous fbf5c51c1c Merge branch 'fix_batch_timesteps' of https://github.com/asagi4/ComfyUI 2 years ago
comfyanonymous 68be24eead Remove some prints. 2 years ago
asagi4 1ea4d84691 Fix timestep ranges when batch_size > 1 2 years ago
comfyanonymous 5379051d16 Fix diffusers VAE loading. 2 years ago
comfyanonymous 727588d076 Fix some new loras. 2 years ago
comfyanonymous 4f9b6f39d1 Fix potential issue with Save Checkpoint. 2 years ago
comfyanonymous 5f75d784a1 Start is now 0.0 and end is now 1.0 for the timestep ranges. 2 years ago
comfyanonymous 7ff14b62f8 ControlNetApplyAdvanced can now define when controlnet gets applied. 2 years ago
comfyanonymous d191c4f9ed Add a ControlNetApplyAdvanced node.
The controlnet can be applied to the positive or negative prompt only by
connecting it correctly.
2 years ago
comfyanonymous 0240946ecf Add a way to set which range of timesteps the cond gets applied to. 2 years ago