199 Commits (f257fc999fde8e9695e4755902c75e7f7192fe2b)

Author SHA1 Message Date
comfyanonymous 5f60ee246e Support loading the sr cascade controlnet. 12 months ago
comfyanonymous 03e6e81629 Set upscale algorithm to bilinear for stable cascade controlnet. 12 months ago
comfyanonymous 03e83bb5d0 Support stable cascade canny controlnet. 12 months ago
comfyanonymous cb7c3a2921 Allow image_only_indicator to be None. 1 year ago
comfyanonymous b3e97fc714 Koala 700M and 1B support.
Use the UNET Loader node to load the unet file to use them.
1 year ago
comfyanonymous e93cdd0ad0 Remove print. 1 year ago
comfyanonymous a7b5eaa7e3 Forgot to commit this. 1 year ago
comfyanonymous 6bcf57ff10 Fix attention masks properly for multiple batches. 1 year ago
comfyanonymous 11e3221f1f fp8 weight support for Stable Cascade. 1 year ago
comfyanonymous f8706546f3 Fix attention mask batch size in some attention functions. 1 year ago
comfyanonymous 3b9969c1c5 Properly fix attention masks in CLIP with batches. 1 year ago
comfyanonymous 805c36ac9c Make Stable Cascade work on old pytorch 2.0 1 year ago
comfyanonymous 667c92814e Stable Cascade Stage B. 1 year ago
comfyanonymous f83109f09b Stable Cascade Stage C. 1 year ago
comfyanonymous 5e06baf112 Stable Cascade Stage A. 1 year ago
comfyanonymous c661a8b118 Don't use numpy for calculating sigmas. 1 year ago
comfyanonymous 89507f8adf Remove some unused imports. 1 year ago
comfyanonymous 2395ae740a Make unclip more deterministic.
Pass a seed argument note that this might make old unclip images different.
1 year ago
comfyanonymous 6a7bc35db8 Use basic attention implementation for small inputs on old pytorch. 1 year ago
comfyanonymous c6951548cf Update optimized_attention_for_device function for new functions that
support masked attention.
1 year ago
comfyanonymous aaa9017302 Add attention mask support to sub quad attention. 1 year ago
comfyanonymous 0c2c9fbdfa Support attention mask in split attention. 1 year ago
comfyanonymous 3ad0191bfb Implement attention mask on xformers. 1 year ago
comfyanonymous 8c6493578b Implement noise augmentation for SD 4X upscale model. 1 year ago
comfyanonymous 79f73a4b33 Remove useless code. 1 year ago
comfyanonymous 61b3f15f8f Fix lowvram mode not working with unCLIP and Revision code. 1 year ago
comfyanonymous d0165d819a Fix SVD lowvram mode. 1 year ago
comfyanonymous 261bcbb0d9 A few missing comfy ops in the VAE. 1 year ago
comfyanonymous a5056cfb1f Remove useless code. 1 year ago
comfyanonymous 77755ab8db Refactor comfy.ops
comfy.ops -> comfy.ops.disable_weight_init

This should make it more clear what they actually do.

Some unused code has also been removed.
1 year ago
comfyanonymous fbdb14d4c4 Cleaner CLIP text encoder implementation.
Use a simple CLIP model implementation instead of the one from
transformers.

This will allow some interesting things that would too hackish to implement
using the transformers implementation.
1 year ago
comfyanonymous 1bbd65ab30 Missed this one. 1 year ago
comfyanonymous 31b0f6f3d8 UNET weights can now be stored in fp8.
--fp8_e4m3fn-unet and --fp8_e5m2-unet are the two different formats
supported by pytorch.
1 year ago
comfyanonymous af365e4dd1 All the unet ops with weights are now handled by comfy.ops 1 year ago
comfyanonymous 39e75862b2 Fix regression from last commit. 1 year ago
comfyanonymous 50dc39d6ec Clean up the extra_options dict for the transformer patches.
Now everything in transformer_options gets put in extra_options.
1 year ago
comfyanonymous 3e5ea74ad3 Make buggy xformers fall back on pytorch attention. 1 year ago
comfyanonymous 871cc20e13 Support SVD img2vid model. 1 year ago
comfyanonymous 72741105a6 Remove useless code. 1 year ago
comfyanonymous 7e3fe3ad28 Make deep shrink behave like it should. 1 year ago
comfyanonymous 7ea6bb038c Print warning when controlnet can't be applied instead of crashing. 1 year ago
comfyanonymous 94cc718e9c Add a way to add patches to the input block. 1 year ago
comfyanonymous 794dd2064d Fix typo. 1 year ago
comfyanonymous a527d0c795 Code refactor. 1 year ago
comfyanonymous 2a23ba0b8c Fix unet ops not entirely on GPU. 1 year ago
comfyanonymous a268a574fa Remove a bunch of useless code.
DDIM is the same as euler with a small difference in the inpaint code.
DDIM uses randn_like but I set a fixed seed instead.

I'm keeping it in because I'm sure if I remove it people are going to
complain.
1 year ago
comfyanonymous c837a173fa Fix some memory issues in sub quad attention. 1 year ago
comfyanonymous 125b03eead Fix some OOM issues with split attention. 1 year ago
comfyanonymous 6ec3f12c6e Support SSD1B model and make it easier to support asymmetric unets. 1 year ago
comfyanonymous a373367b0c Fix some OOM issues with split and sub quad attention. 1 year ago
comfyanonymous 8b65f5de54 attention_basic now works with hypertile. 1 year ago
comfyanonymous e6bc42df46 Make sub_quad and split work with hypertile. 1 year ago
comfyanonymous 9906e3efe3 Make xformers work with hypertile. 1 year ago
comfyanonymous d44a2de49f Make VAE code closer to sgm. 1 year ago
comfyanonymous 23680a9155 Refactor the attention stuff in the VAE. 1 year ago
comfyanonymous bb064c9796 Add a separate optimized_attention_masked function. 1 year ago
comfyanonymous 9a55dadb4c Refactor code so model can be a dtype other than fp32 or fp16. 1 year ago
comfyanonymous 88733c997f pytorch_attention_enabled can now return True when xformers is enabled. 1 year ago
comfyanonymous ac7d8cfa87 Allow attn_mask in attention_pytorch. 1 year ago
comfyanonymous 1a4bd9e9a6 Refactor the attention functions.
There's no reason for the whole CrossAttention object to be repeated when
only the operation in the middle changes.
1 year ago
comfyanonymous fff491b032 Model patches can now know which batch is positive and negative. 1 year ago
comfyanonymous 446caf711c Sampling code refactor. 1 year ago
comfyanonymous afa2399f79 Add a way to set output block patches to modify the h and hsp. 1 year ago
comfyanonymous 94e4fe39d8 This isn't used anywhere. 1 year ago
comfyanonymous 1938f5c5fe Add a force argument to soft_empty_cache to force a cache empty. 1 year ago
Simon Lui 2da73b7073 Revert changes in comfy/ldm/modules/diffusionmodules/util.py, which is unused. 1 year ago
Simon Lui 4a0c4ce4ef Some fixes to generalize CUDA specific functionality to Intel or other GPUs. 1 year ago
comfyanonymous 0e3b641172 Remove xformers related print. 2 years ago
comfyanonymous bed116a1f9 Remove optimization that caused border. 2 years ago
comfyanonymous 1c794a2161 Fallback to slice attention if xformers doesn't support the operation. 2 years ago
comfyanonymous d935ba50c4 Make --bf16-vae work on torch 2.0 2 years ago
comfyanonymous cf5ae46928 Controlnet/t2iadapter cleanup. 2 years ago
comfyanonymous b80c3276dc Fix issue with gligen. 2 years ago
comfyanonymous d6e4b342e6 Support for Control Loras.
Control loras are controlnets where some of the weights are stored in
"lora" format: an up and a down low rank matrice that when multiplied
together and added to the unet weight give the controlnet weight.

This allows a much smaller memory footprint depending on the rank of the
matrices.

These controlnets are used just like regular ones.
2 years ago
comfyanonymous 2b13939044 Remove some useless code. 2 years ago
comfyanonymous 95d796fc85 Faster VAE loading. 2 years ago
comfyanonymous 4b957a0010 Initialize the unet directly on the target device. 2 years ago
comfyanonymous 9ba440995a It's actually possible to torch.compile the unet now. 2 years ago
comfyanonymous 3ded1a3a04 Refactor of sampler code to deal more easily with different model types. 2 years ago
comfyanonymous ddc6f12ad5 Disable autocast in unet for increased speed. 2 years ago
comfyanonymous 103c487a89 Cleanup. 2 years ago
comfyanonymous c71a7e6b20 Fix ddim + inpainting not working. 2 years ago
comfyanonymous 78d8035f73 Fix bug with controlnet. 2 years ago
comfyanonymous 05676942b7 Add some more transformer hooks and move tomesd to comfy_extras.
Tomesd now uses q instead of x to decide which tokens to merge because
it seems to give better results.
2 years ago
comfyanonymous fa28d7334b Remove useless code. 2 years ago
comfyanonymous f87ec10a97 Support base SDXL and SDXL refiner models.
Large refactor of the model detection and loading code.
2 years ago
comfyanonymous 9fccf4aa03 Add original_shape parameter to transformer patch extra_options. 2 years ago
comfyanonymous 8883cb0f67 Add a way to set patches that modify the attn2 output.
Change the transformer patches function format to be more future proof.
2 years ago
comfyanonymous 45be2e92c1 Fix DDIM v-prediction. 2 years ago
comfyanonymous ae43f09ef7 All the unet weights should now be initialized with the right dtype. 2 years ago
comfyanonymous 7bf89ba923 Initialize more unet weights as the right dtype. 2 years ago
comfyanonymous e21d9ad445 Initialize transformer unet block weights in right dtype at the start. 2 years ago
comfyanonymous 21f04fe632 Disable default weight values in unet conv2d for faster loading. 2 years ago
comfyanonymous 9d54066ebc This isn't needed for inference. 2 years ago
comfyanonymous 6971646b8b Speed up model loading a bit.
Default pytorch Linear initializes the weights which is useless and slow.
2 years ago
comfyanonymous 274dff3257 Remove more useless files. 2 years ago
comfyanonymous f0a2b81cd0 Cleanup: Remove a bunch of useless files. 2 years ago
comfyanonymous b8636a44aa Make scaled_dot_product switch to sliced attention on OOM. 2 years ago
comfyanonymous 797c4e8d3b Simplify and improve some vae attention code. 2 years ago
BlenderNeko d9e088ddfd minor changes for tiled sampler 2 years ago