672 Commits (079dbf919874e6fce170d316e409366bd409cfb9)

Author SHA1 Message Date
comfyanonymous 036f88c621 Refactor to make it easier to add custom conds to models. 1 year ago
comfyanonymous 3fce8881ca Sampling code refactor to make it easier to add more conds. 1 year ago
comfyanonymous 8594c8be4d Empty the cache when torch cache is more than 25% free mem. 1 year ago
comfyanonymous 8b65f5de54 attention_basic now works with hypertile. 1 year ago
comfyanonymous e6bc42df46 Make sub_quad and split work with hypertile. 1 year ago
comfyanonymous a0690f9df9 Fix t2i adapter issue. 1 year ago
comfyanonymous 9906e3efe3 Make xformers work with hypertile. 1 year ago
comfyanonymous 4185324a1d Fix uni_pc sampler math. This changes the images this sampler produces. 1 year ago
comfyanonymous e6962120c6 Make sure cond_concat is on the right device. 1 year ago
comfyanonymous 45c972aba8 Refactor cond_concat into conditioning. 1 year ago
comfyanonymous 430a8334c5 Fix some potential issues. 1 year ago
comfyanonymous 782a24fce6 Refactor cond_concat into model object. 1 year ago
comfyanonymous 0d45a565da Fix memory issue related to control loras.
The cleanup function was not getting called.
1 year ago
comfyanonymous d44a2de49f Make VAE code closer to sgm. 1 year ago
comfyanonymous 23680a9155 Refactor the attention stuff in the VAE. 1 year ago
comfyanonymous c8013f73e5 Add some Quadro cards to the list of cards with broken fp16. 1 year ago
comfyanonymous bb064c9796 Add a separate optimized_attention_masked function. 1 year ago
comfyanonymous fd4c5f07e7 Add a --bf16-unet to test running the unet in bf16. 1 year ago
comfyanonymous 9a55dadb4c Refactor code so model can be a dtype other than fp32 or fp16. 1 year ago
comfyanonymous 88733c997f pytorch_attention_enabled can now return True when xformers is enabled. 1 year ago
comfyanonymous 20d3852aa1 Pull some small changes from the other repo. 1 year ago
comfyanonymous ac7d8cfa87 Allow attn_mask in attention_pytorch. 1 year ago
comfyanonymous 1a4bd9e9a6 Refactor the attention functions.
There's no reason for the whole CrossAttention object to be repeated when
only the operation in the middle changes.
1 year ago
comfyanonymous 8cc75c64ff Let unet wrapper functions have .to attributes. 1 year ago
comfyanonymous 5e885bd9c8 Cleanup. 1 year ago
comfyanonymous 851bb87ca9 Merge branch 'taesd_safetensors' of https://github.com/mochiya98/ComfyUI 1 year ago
Yukimasa Funaoka 9eb621c95a
Supports TAESD models in safetensors format 1 year ago
comfyanonymous d1a0abd40b Merge branch 'input-directory' of https://github.com/jn-jairo/ComfyUI 1 year ago
comfyanonymous 72188dffc3 load_checkpoint_guess_config can now optionally output the model. 1 year ago
Jairo Correa 63e5fd1790 Option to input directory 1 year ago
City 9bfec2bdbf Fix quality loss due to low precision 1 year ago
badayvedat 0f17993d05 fix: typo in extra sampler 1 year ago
comfyanonymous 66756de100 Add SamplerDPMPP_2M_SDE node. 1 year ago
comfyanonymous 71713888c4 Print missing VAE keys. 1 year ago
comfyanonymous d234ca558a Add missing samplers to KSamplerSelect. 1 year ago
comfyanonymous 1adcc4c3a2 Add a SamplerCustom Node.
This node takes a list of sigmas and a sampler object as input.

This lets people easily implement custom schedulers and samplers as nodes.

More nodes will be added to it in the future.
1 year ago
comfyanonymous bf3fc2f1b7 Refactor sampling related code. 1 year ago
comfyanonymous fff491b032 Model patches can now know which batch is positive and negative. 1 year ago
comfyanonymous 1d6dd83184 Scheduler code refactor. 1 year ago
comfyanonymous 446caf711c Sampling code refactor. 1 year ago
comfyanonymous 76cdc809bf Support more controlnet models. 1 year ago
comfyanonymous ae87543653 Merge branch 'cast_intel' of https://github.com/simonlui/ComfyUI 1 year ago
Simon Lui eec449ca8e Allow Intel GPUs to LoRA cast on GPU since it supports BF16 natively. 1 year ago
comfyanonymous afa2399f79 Add a way to set output block patches to modify the h and hsp. 1 year ago
comfyanonymous 492db2de8d Allow having a different pooled output for each image in a batch. 1 year ago
comfyanonymous 1cdfb3dba4 Only do the cast on the device if the device supports it. 1 year ago
comfyanonymous 7c9a92f552 Don't depend on torchvision. 1 year ago
MoonRide303 2b6b178173 Added support for lanczos scaling 1 year ago
comfyanonymous b92bf8196e Do lora cast on GPU instead of CPU for higher performance. 1 year ago
comfyanonymous 321c5fa295 Enable pytorch attention by default on xpu. 1 year ago
comfyanonymous 61b1f67734 Support models without previews. 1 year ago
comfyanonymous 43d4935a1d Add cond_or_uncond array to transformer_options so hooks can check what is
cond and what is uncond.
1 year ago
comfyanonymous 415abb275f Add DDPM sampler. 1 year ago
comfyanonymous 94e4fe39d8 This isn't used anywhere. 1 year ago
comfyanonymous 44361f6344 Support for text encoder models that need attention_mask. 1 year ago
comfyanonymous 0d8f376446 Set last layer on SD2.x models uses the proper indexes now.
Before I had made the last layer the penultimate layer because some
checkpoints don't have them but it's not consistent with the others models.

TLDR: for SD2.x models only: CLIPSetLastLayer -1 is now -2.
1 year ago
comfyanonymous 0966d3ce82 Don't run text encoders on xpu because there are issues. 1 year ago
comfyanonymous 3039b08eb1 Only parse command line args when main.py is called. 1 year ago
comfyanonymous ed58730658 Don't leave very large hidden states in the clip vision output. 1 year ago
comfyanonymous fb3b728203 Fix issue where autocast fp32 CLIP gave different results from regular. 1 year ago
comfyanonymous 7d401ed1d0 Add ldm format support to UNETLoader. 1 year ago
comfyanonymous e85be36bd2 Add a penultimate_hidden_states to the clip vision output. 1 year ago
comfyanonymous 1e6b67101c Support diffusers format t2i adapters. 1 year ago
comfyanonymous 326577d04c Allow cancelling of everything with a progress bar. 1 year ago
comfyanonymous f88f7f413a Add a ConditioningSetAreaPercentage node. 1 year ago
comfyanonymous 1938f5c5fe Add a force argument to soft_empty_cache to force a cache empty. 1 year ago
comfyanonymous 7746bdf7b0 Merge branch 'generalize_fixes' of https://github.com/simonlui/ComfyUI 1 year ago
Simon Lui 2da73b7073 Revert changes in comfy/ldm/modules/diffusionmodules/util.py, which is unused. 1 year ago
comfyanonymous a74c5dbf37 Move some functions to utils.py 1 year ago
Simon Lui 4a0c4ce4ef Some fixes to generalize CUDA specific functionality to Intel or other GPUs. 1 year ago
comfyanonymous 77a176f9e0 Use common function to reshape batch to. 2 years ago
comfyanonymous 7931ff0fd9 Support SDXL inpaint models. 2 years ago
comfyanonymous 0e3b641172 Remove xformers related print. 2 years ago
comfyanonymous 5c363a9d86 Fix controlnet bug. 2 years ago
comfyanonymous cfe1c54de8 Fix controlnet issue. 2 years ago
comfyanonymous 1c012d69af It doesn't make sense for c_crossattn and c_concat to be lists. 2 years ago
comfyanonymous 7e941f9f24 Clean up DiffusersLoader node. 2 years ago
Simon Lui 18617967e5
Fix error message in model_patcher.py
Found while tinkering.
2 years ago
comfyanonymous fe4c07400c Fix "Load Checkpoint with config" node. 2 years ago
comfyanonymous f2f5e5dcbb Support SDXL t2i adapters with 3 channel input. 2 years ago
comfyanonymous 15adc3699f Move beta_schedule to model_config and allow disabling unet creation. 2 years ago
comfyanonymous bed116a1f9 Remove optimization that caused border. 2 years ago
comfyanonymous 65cae62c71 No need to check filename extensions to detect shuffle controlnet. 2 years ago
comfyanonymous 4e89b2c25a Put clip vision outputs on the CPU. 2 years ago
comfyanonymous a094b45c93 Load clipvision model to GPU for faster performance. 2 years ago
comfyanonymous 1300a1bb4c Text encoder should initially load on the offload_device not the regular. 2 years ago
comfyanonymous f92074b84f Move ModelPatcher to model_patcher.py 2 years ago
comfyanonymous 4798cf5a62 Implement loras with norm keys. 2 years ago
comfyanonymous b8c7c770d3 Enable bf16-vae by default on ampere and up. 2 years ago
comfyanonymous 1c794a2161 Fallback to slice attention if xformers doesn't support the operation. 2 years ago
comfyanonymous d935ba50c4 Make --bf16-vae work on torch 2.0 2 years ago
comfyanonymous a57b0c797b Fix lowvram model merging. 2 years ago
comfyanonymous f72780a7e3 The new smart memory management makes this unnecessary. 2 years ago
comfyanonymous c77f02e1c6 Move controlnet code to comfy/controlnet.py 2 years ago
comfyanonymous 15a7716fa6 Move lora code to comfy/lora.py 2 years ago
comfyanonymous ec96f6d03a Move text_projection to base clip model. 2 years ago
comfyanonymous 30eb92c3cb Code cleanups. 2 years ago
comfyanonymous 51dde87e97 Try to free enough vram for control lora inference. 2 years ago
comfyanonymous e3d0a9a490 Fix potential issue with text projection matrix multiplication. 2 years ago
comfyanonymous cc44ade79e Always shift text encoder to GPU when the device supports fp16. 2 years ago
comfyanonymous a6ef08a46a Even with forced fp16 the cpu device should never use it. 2 years ago
comfyanonymous 00c0b2c507 Initialize text encoder to target dtype. 2 years ago
comfyanonymous f081017c1a Save memory by storing text encoder weights in fp16 in most situations.
Do inference in fp32 to make sure quality stays the exact same.
2 years ago
comfyanonymous afcb9cb1df All resolutions now work with t2i adapter for SDXL. 2 years ago
comfyanonymous 85fde89d7f T2I adapter SDXL. 2 years ago
comfyanonymous cf5ae46928 Controlnet/t2iadapter cleanup. 2 years ago
comfyanonymous 763b0cf024 Fix control lora not working in fp32. 2 years ago
comfyanonymous 199d73364a Fix ControlLora on lowvram. 2 years ago
comfyanonymous d08e53de2e Remove autocast from controlnet code. 2 years ago
comfyanonymous 0d7b0a4dc7 Small cleanups. 2 years ago
Simon Lui 9225465975 Further tuning and fix mem_free_total. 2 years ago
Simon Lui 2c096e4260 Add ipex optimize and other enhancements for Intel GPUs based on recent memory changes. 2 years ago
comfyanonymous e9469e732d --disable-smart-memory now disables loading model directly to vram. 2 years ago
comfyanonymous c9b562aed1 Free more memory before VAE encode/decode. 2 years ago
comfyanonymous b80c3276dc Fix issue with gligen. 2 years ago
comfyanonymous d6e4b342e6 Support for Control Loras.
Control loras are controlnets where some of the weights are stored in
"lora" format: an up and a down low rank matrice that when multiplied
together and added to the unet weight give the controlnet weight.

This allows a much smaller memory footprint depending on the rank of the
matrices.

These controlnets are used just like regular ones.
2 years ago
comfyanonymous 39ac856a33 ReVision support: unclip nodes can now be used with SDXL. 2 years ago
comfyanonymous 76d53c4622 Add support for clip g vision model to CLIPVisionLoader. 2 years ago
Alexopus e59fe0537a
Fix referenced before assignment
For https://github.com/BlenderNeko/ComfyUI_TiledKSampler/issues/13
2 years ago
comfyanonymous be9c5e25bc Fix issue with not freeing enough memory when sampling. 2 years ago
comfyanonymous ac0758a1a4 Fix bug with lowvram and controlnet advanced node. 2 years ago
comfyanonymous c28db1f315 Fix potential issues with patching models when saving checkpoints. 2 years ago
comfyanonymous 3aee33b54e Add --disable-smart-memory for those that want the old behaviour. 2 years ago
comfyanonymous 2be2742711 Fix issue with regular torch version. 2 years ago
comfyanonymous 89a0767abf Smarter memory management.
Try to keep models on the vram when possible.

Better lowvram mode for controlnets.
2 years ago
comfyanonymous 2c97c30256 Support small diffusers controlnet so both types are now supported. 2 years ago
comfyanonymous 53f326a3d8 Support diffusers mini controlnets. 2 years ago
comfyanonymous 58f0c616ed Fix clip vision issue with old transformers versions. 2 years ago
comfyanonymous ae270f79bc Fix potential issue with batch size and clip vision. 2 years ago
comfyanonymous a2ce9655ca Refactor unclip code. 2 years ago
comfyanonymous 9cc12c833d CLIPVisionEncode can now encode multiple images. 2 years ago
comfyanonymous 0cb6dac943 Remove 3m from PR #1213 because of some small issues. 2 years ago
comfyanonymous e244b2df83 Add sgm_uniform scheduler that acts like the default one in sgm. 2 years ago
comfyanonymous 58c7da3665 Gpu variant of dpmpp_3m_sde. Note: use 3m with exponential or karras. 2 years ago
comfyanonymous ba319a34e4 Merge branch 'dpmpp3m' of https://github.com/FizzleDorf/ComfyUI 2 years ago
FizzleDorf 3cfad03a68 dpmpp 3m + dpmpp 3m sde added 2 years ago
comfyanonymous 585a062910 Print unet config when model isn't detected. 2 years ago
comfyanonymous c8a23ce9e8 Support for yet another lora type based on diffusers. 2 years ago
comfyanonymous 2bc12d3d22 Add --temp-directory argument to set temp directory. 2 years ago
comfyanonymous c20583286f Support diffuser text encoder loras. 2 years ago
comfyanonymous cf10c5592c Disable calculating uncond when CFG is 1.0 2 years ago
comfyanonymous 1f0f4cc0bd Add argument to disable auto launching the browser. 2 years ago
comfyanonymous d8e58f0a7e Detect hint_channels from controlnet. 2 years ago
comfyanonymous c5d7593ccf Support loras in diffusers format. 2 years ago
comfyanonymous 1ce0d8ad68 Add CMP 30HX card to the nvidia_16_series list. 2 years ago
comfyanonymous c99d8002f8 Make sure the pooled output stays at the EOS token with added embeddings. 2 years ago
comfyanonymous 4a77fcd6ab Only shift text encoder to vram when CPU cores are under 8. 2 years ago
comfyanonymous 3cd31d0e24 Lower CPU thread check for running the text encoder on the CPU vs GPU. 2 years ago
comfyanonymous 2b13939044 Remove some useless code. 2 years ago
comfyanonymous 95d796fc85 Faster VAE loading. 2 years ago
comfyanonymous 4b957a0010 Initialize the unet directly on the target device. 2 years ago
comfyanonymous c910b4a01c Remove unused code and torchdiffeq dependency. 2 years ago
comfyanonymous 1141029a4a Add --disable-metadata argument to disable saving metadata in files. 2 years ago
comfyanonymous fbf5c51c1c Merge branch 'fix_batch_timesteps' of https://github.com/asagi4/ComfyUI 2 years ago
comfyanonymous 68be24eead Remove some prints. 2 years ago
asagi4 1ea4d84691 Fix timestep ranges when batch_size > 1 2 years ago
comfyanonymous 5379051d16 Fix diffusers VAE loading. 2 years ago
comfyanonymous 727588d076 Fix some new loras. 2 years ago
comfyanonymous 4f9b6f39d1 Fix potential issue with Save Checkpoint. 2 years ago
comfyanonymous 5f75d784a1 Start is now 0.0 and end is now 1.0 for the timestep ranges. 2 years ago
comfyanonymous 7ff14b62f8 ControlNetApplyAdvanced can now define when controlnet gets applied. 2 years ago
comfyanonymous d191c4f9ed Add a ControlNetApplyAdvanced node.
The controlnet can be applied to the positive or negative prompt only by
connecting it correctly.
2 years ago
comfyanonymous 0240946ecf Add a way to set which range of timesteps the cond gets applied to. 2 years ago
comfyanonymous 22f29d66ca Try to fix memory issue with lora. 2 years ago
comfyanonymous 67be7eb81d Nodes can now patch the unet function. 2 years ago
comfyanonymous 12a6e93171 Del the right object when applying lora. 2 years ago
comfyanonymous 78e7958d17 Support controlnet in diffusers format. 2 years ago
comfyanonymous 09386a3697 Fix issue with lora in some cases when combined with model merging. 2 years ago
comfyanonymous 58b2364f58 Properly support SDXL diffusers unet with UNETLoader node. 2 years ago
comfyanonymous 0115018695 Print errors and continue when lora weights are not compatible. 2 years ago
comfyanonymous 4760c29380 Merge branch 'fix-AttributeError-module-'torch'-has-no-attribute-'mps'' of https://github.com/KarryCharon/ComfyUI 2 years ago
comfyanonymous 0b284f650b Fix typo. 2 years ago
comfyanonymous e032ca6138 Fix ddim issue with older torch versions. 2 years ago
comfyanonymous 18885f803a Add MX450 and MX550 to list of cards with broken fp16. 2 years ago
comfyanonymous 9ba440995a It's actually possible to torch.compile the unet now. 2 years ago
comfyanonymous 51d5477579 Add key to indicate checkpoint is v_prediction when saving. 2 years ago
comfyanonymous ff6b047a74 Fix device print on old torch version. 2 years ago
comfyanonymous 9871a15cf9 Enable --cuda-malloc by default on torch 2.0 and up.
Add --disable-cuda-malloc to disable it.
2 years ago
comfyanonymous 55d0fca9fa --windows-standalone-build now enables --cuda-malloc 2 years ago
comfyanonymous 1679abd86d Add a command line argument to enable backend:cudaMallocAsync 2 years ago
comfyanonymous 3a150bad15 Only calculate randn in some samplers when it's actually being used. 2 years ago
comfyanonymous ee8f8ee07f Fix regression with ddim and uni_pc when batch size > 1. 2 years ago
comfyanonymous 3ded1a3a04 Refactor of sampler code to deal more easily with different model types. 2 years ago
comfyanonymous 5f57362613 Lower lora ram usage when in normal vram mode. 2 years ago
comfyanonymous 490771b7f4 Speed up lora loading a bit. 2 years ago
comfyanonymous 50b1180dde Fix CLIPSetLastLayer not reverting when removed. 2 years ago
comfyanonymous 6fb084f39d Reduce floating point rounding errors in loras. 2 years ago
comfyanonymous 91ed2815d5 Add a node to merge CLIP models. 2 years ago
comfyanonymous b2f03164c7 Prevent the clip_g position_ids key from being saved in the checkpoint.
This is to make it match the official checkpoint.
2 years ago
comfyanonymous 46dc050c9f Fix potential tensors being on different devices issues. 2 years ago
KarryCharon 3e2309f149 fix mps miss import 2 years ago
comfyanonymous 606a537090 Support SDXL embedding format with 2 CLIP. 2 years ago
comfyanonymous 6ad0a6d7e2 Don't patch weights when multiplier is zero. 2 years ago
comfyanonymous d5323d16e0 latent2rgb matrix for SDXL. 2 years ago
comfyanonymous 0ae81c03bb Empty cache after model unloading for normal vram and lower. 2 years ago
comfyanonymous d3f5998218 Support loading clip_g from diffusers in CLIP Loader nodes. 2 years ago
comfyanonymous a9a4ba7574 Fix merging not working when model2 of model merge node was a merge. 2 years ago
comfyanonymous bb5fbd29e9 Merge branch 'condmask-fix' of https://github.com/vmedea/ComfyUI 2 years ago
comfyanonymous e7bee85df8 Add arguments to run the VAE in fp16 or bf16 for testing. 2 years ago
comfyanonymous 608fcc2591 Fix bug with weights when prompt is long. 2 years ago