comfyanonymous
261bcbb0d9
A few missing comfy ops in the VAE.
1 year ago
comfyanonymous
a5056cfb1f
Remove useless code.
1 year ago
comfyanonymous
77755ab8db
Refactor comfy.ops
...
comfy.ops -> comfy.ops.disable_weight_init
This should make it more clear what they actually do.
Some unused code has also been removed.
1 year ago
comfyanonymous
fbdb14d4c4
Cleaner CLIP text encoder implementation.
...
Use a simple CLIP model implementation instead of the one from
transformers.
This will allow some interesting things that would too hackish to implement
using the transformers implementation.
1 year ago
comfyanonymous
1bbd65ab30
Missed this one.
1 year ago
comfyanonymous
31b0f6f3d8
UNET weights can now be stored in fp8.
...
--fp8_e4m3fn-unet and --fp8_e5m2-unet are the two different formats
supported by pytorch.
1 year ago
comfyanonymous
af365e4dd1
All the unet ops with weights are now handled by comfy.ops
1 year ago
comfyanonymous
39e75862b2
Fix regression from last commit.
1 year ago
comfyanonymous
50dc39d6ec
Clean up the extra_options dict for the transformer patches.
...
Now everything in transformer_options gets put in extra_options.
1 year ago
comfyanonymous
3e5ea74ad3
Make buggy xformers fall back on pytorch attention.
1 year ago
comfyanonymous
871cc20e13
Support SVD img2vid model.
1 year ago
comfyanonymous
72741105a6
Remove useless code.
1 year ago
comfyanonymous
7e3fe3ad28
Make deep shrink behave like it should.
1 year ago
comfyanonymous
7ea6bb038c
Print warning when controlnet can't be applied instead of crashing.
1 year ago
comfyanonymous
94cc718e9c
Add a way to add patches to the input block.
1 year ago
comfyanonymous
794dd2064d
Fix typo.
1 year ago
comfyanonymous
a527d0c795
Code refactor.
1 year ago
comfyanonymous
2a23ba0b8c
Fix unet ops not entirely on GPU.
1 year ago
comfyanonymous
c837a173fa
Fix some memory issues in sub quad attention.
1 year ago
comfyanonymous
125b03eead
Fix some OOM issues with split attention.
1 year ago
comfyanonymous
6ec3f12c6e
Support SSD1B model and make it easier to support asymmetric unets.
1 year ago
comfyanonymous
a373367b0c
Fix some OOM issues with split and sub quad attention.
1 year ago
comfyanonymous
8b65f5de54
attention_basic now works with hypertile.
1 year ago
comfyanonymous
e6bc42df46
Make sub_quad and split work with hypertile.
1 year ago
comfyanonymous
9906e3efe3
Make xformers work with hypertile.
1 year ago
comfyanonymous
d44a2de49f
Make VAE code closer to sgm.
1 year ago
comfyanonymous
23680a9155
Refactor the attention stuff in the VAE.
1 year ago
comfyanonymous
bb064c9796
Add a separate optimized_attention_masked function.
1 year ago
comfyanonymous
9a55dadb4c
Refactor code so model can be a dtype other than fp32 or fp16.
1 year ago
comfyanonymous
88733c997f
pytorch_attention_enabled can now return True when xformers is enabled.
1 year ago
comfyanonymous
ac7d8cfa87
Allow attn_mask in attention_pytorch.
1 year ago
comfyanonymous
1a4bd9e9a6
Refactor the attention functions.
...
There's no reason for the whole CrossAttention object to be repeated when
only the operation in the middle changes.
1 year ago
comfyanonymous
fff491b032
Model patches can now know which batch is positive and negative.
1 year ago
comfyanonymous
afa2399f79
Add a way to set output block patches to modify the h and hsp.
1 year ago
comfyanonymous
1938f5c5fe
Add a force argument to soft_empty_cache to force a cache empty.
1 year ago
Simon Lui
2da73b7073
Revert changes in comfy/ldm/modules/diffusionmodules/util.py, which is unused.
1 year ago
Simon Lui
4a0c4ce4ef
Some fixes to generalize CUDA specific functionality to Intel or other GPUs.
1 year ago
comfyanonymous
0e3b641172
Remove xformers related print.
2 years ago
comfyanonymous
bed116a1f9
Remove optimization that caused border.
2 years ago
comfyanonymous
1c794a2161
Fallback to slice attention if xformers doesn't support the operation.
2 years ago
comfyanonymous
d935ba50c4
Make --bf16-vae work on torch 2.0
2 years ago
comfyanonymous
cf5ae46928
Controlnet/t2iadapter cleanup.
2 years ago
comfyanonymous
b80c3276dc
Fix issue with gligen.
2 years ago
comfyanonymous
d6e4b342e6
Support for Control Loras.
...
Control loras are controlnets where some of the weights are stored in
"lora" format: an up and a down low rank matrice that when multiplied
together and added to the unet weight give the controlnet weight.
This allows a much smaller memory footprint depending on the rank of the
matrices.
These controlnets are used just like regular ones.
2 years ago
comfyanonymous
2b13939044
Remove some useless code.
2 years ago
comfyanonymous
95d796fc85
Faster VAE loading.
2 years ago
comfyanonymous
4b957a0010
Initialize the unet directly on the target device.
2 years ago
comfyanonymous
9ba440995a
It's actually possible to torch.compile the unet now.
2 years ago
comfyanonymous
ddc6f12ad5
Disable autocast in unet for increased speed.
2 years ago
comfyanonymous
103c487a89
Cleanup.
2 years ago