641 Commits (a252963f956a7d76344e3f0ce24b1047480a25af)

Author SHA1 Message Date
comfyanonymous a252963f95 --disable-smart-memory now unloads everything like it did originally. 1 year ago
comfyanonymous 36a7953142 Greatly improve lowvram sampling speed by getting rid of accelerate.
Let me know if this breaks anything.
1 year ago
comfyanonymous 261bcbb0d9 A few missing comfy ops in the VAE. 1 year ago
comfyanonymous 9a7619b72d Fix regression with inpaint model. 1 year ago
comfyanonymous 571ea8cdcc Fix SAG not working with cfg 1.0 1 year ago
comfyanonymous 8cf1daa108 Fix SDXL area composition sometimes not using the right pooled output. 1 year ago
comfyanonymous 2258f85159 Support stable zero 123 model.
To use it use the ImageOnlyCheckpointLoader to load the checkpoint and
the new Stable_Zero123 node.
1 year ago
comfyanonymous 2f9d6a97ec Add --deterministic option to make pytorch use deterministic algorithms. 1 year ago
comfyanonymous e45d920ae3 Don't resize clip vision image when the size is already good. 1 year ago
comfyanonymous 13e6d5366e Switch clip vision to manual cast.
Make it use the same dtype as the text encoder.
1 year ago
comfyanonymous 719fa0866f Set clip vision model in eval mode so it works without inference mode. 1 year ago
Hari 574363a8a6 Implement Perp-Neg 1 year ago
comfyanonymous a5056cfb1f Remove useless code. 1 year ago
comfyanonymous 329c571993 Improve code legibility. 1 year ago
comfyanonymous 6c5990f7db Fix cfg being calculated more than once if sampler_cfg_function. 1 year ago
comfyanonymous ba04a87d10 Refactor and improve the sag node.
Moved all the sag related code to comfy_extras/nodes_sag.py
1 year ago
Rafie Walker 6761233e9d
Implement Self-Attention Guidance (#2201)
* First SAG test

* need to put extra options on the model instead of patcher

* no errors and results seem not-broken

* Use @ashen-uncensored formula, which works better!!!

* Fix a crash when using weird resolutions. Remove an unnecessary UNet call

* Improve comments, optimize memory in blur routine

* SAG works with sampler_cfg_function
1 year ago
comfyanonymous b454a67bb9 Support segmind vega model. 1 year ago
comfyanonymous 824e4935f5 Add dtype parameter to VAE object. 1 year ago
comfyanonymous 32b7e7e769 Add manual cast to controlnet. 1 year ago
comfyanonymous 3152023fbc Use inference dtype for unet memory usage estimation. 1 year ago
comfyanonymous 77755ab8db Refactor comfy.ops
comfy.ops -> comfy.ops.disable_weight_init

This should make it more clear what they actually do.

Some unused code has also been removed.
1 year ago
comfyanonymous b0aab1e4ea Add an option --fp16-unet to force using fp16 for the unet. 1 year ago
comfyanonymous ba07cb748e Use faster manual cast for fp8 in unet. 1 year ago
comfyanonymous 57926635e8 Switch text encoder to manual cast.
Use fp16 text encoder weights for CPU inference to lower memory usage.
1 year ago
comfyanonymous 340177e6e8 Disable non blocking on mps. 1 year ago
comfyanonymous 614b7e731f Implement GLora. 1 year ago
comfyanonymous cb63e230b4 Make lora code a bit cleaner. 1 year ago
comfyanonymous 174eba8e95 Use own clip vision model implementation. 1 year ago
comfyanonymous 97015b6b38 Cleanup. 1 year ago
comfyanonymous a4ec54a40d Add linear_start and linear_end to model_config.sampling_settings 1 year ago
comfyanonymous 9ac0b487ac Make --gpu-only put intermediate values in GPU memory instead of cpu. 1 year ago
comfyanonymous efb704c758 Support attention masking in CLIP implementation. 1 year ago
comfyanonymous fbdb14d4c4 Cleaner CLIP text encoder implementation.
Use a simple CLIP model implementation instead of the one from
transformers.

This will allow some interesting things that would too hackish to implement
using the transformers implementation.
1 year ago
comfyanonymous 2db86b4676 Slightly faster lora applying. 1 year ago
comfyanonymous 1bbd65ab30 Missed this one. 1 year ago
comfyanonymous 9b655d4fd7 Fix memory issue with control loras. 1 year ago
comfyanonymous 26b1c0a771 Fix control lora on fp8. 1 year ago
comfyanonymous be3468ddd5 Less useless downcasting. 1 year ago
comfyanonymous ca82ade765 Use .itemsize to get dtype size for fp8. 1 year ago
comfyanonymous 31b0f6f3d8 UNET weights can now be stored in fp8.
--fp8_e4m3fn-unet and --fp8_e5m2-unet are the two different formats
supported by pytorch.
1 year ago
comfyanonymous af365e4dd1 All the unet ops with weights are now handled by comfy.ops 1 year ago
comfyanonymous 61a123a1e0 A different way of handling multiple images passed to SVD.
Previously when a list of 3 images [0, 1, 2] was used for a 6 frame video
they were concated like this:
[0, 1, 2, 0, 1, 2]

now they are concated like this:
[0, 0, 1, 1, 2, 2]
1 year ago
comfyanonymous c97be4db91 Support SD2.1 turbo checkpoint. 1 year ago
comfyanonymous 983ebc5792 Use smart model management for VAE to decrease latency. 1 year ago
comfyanonymous c45d1b9b67 Add a function to load a unet from a state dict. 1 year ago
comfyanonymous f30b992b18 .sigma and .timestep now return tensors on the same device as the input. 1 year ago
comfyanonymous 13fdee6abf Try to free memory for both cond+uncond before inference. 1 year ago
comfyanonymous be71bb5e13 Tweak memory inference calculations a bit. 1 year ago
comfyanonymous 39e75862b2 Fix regression from last commit. 1 year ago