comfyanonymous
a252963f95
--disable-smart-memory now unloads everything like it did originally.
1 year ago
comfyanonymous
36a7953142
Greatly improve lowvram sampling speed by getting rid of accelerate.
...
Let me know if this breaks anything.
1 year ago
comfyanonymous
261bcbb0d9
A few missing comfy ops in the VAE.
1 year ago
comfyanonymous
9a7619b72d
Fix regression with inpaint model.
1 year ago
comfyanonymous
571ea8cdcc
Fix SAG not working with cfg 1.0
1 year ago
comfyanonymous
8cf1daa108
Fix SDXL area composition sometimes not using the right pooled output.
1 year ago
comfyanonymous
2258f85159
Support stable zero 123 model.
...
To use it use the ImageOnlyCheckpointLoader to load the checkpoint and
the new Stable_Zero123 node.
1 year ago
comfyanonymous
2f9d6a97ec
Add --deterministic option to make pytorch use deterministic algorithms.
1 year ago
comfyanonymous
e45d920ae3
Don't resize clip vision image when the size is already good.
1 year ago
comfyanonymous
13e6d5366e
Switch clip vision to manual cast.
...
Make it use the same dtype as the text encoder.
1 year ago
comfyanonymous
719fa0866f
Set clip vision model in eval mode so it works without inference mode.
1 year ago
Hari
574363a8a6
Implement Perp-Neg
1 year ago
comfyanonymous
a5056cfb1f
Remove useless code.
1 year ago
comfyanonymous
329c571993
Improve code legibility.
1 year ago
comfyanonymous
6c5990f7db
Fix cfg being calculated more than once if sampler_cfg_function.
1 year ago
comfyanonymous
ba04a87d10
Refactor and improve the sag node.
...
Moved all the sag related code to comfy_extras/nodes_sag.py
1 year ago
Rafie Walker
6761233e9d
Implement Self-Attention Guidance ( #2201 )
...
* First SAG test
* need to put extra options on the model instead of patcher
* no errors and results seem not-broken
* Use @ashen-uncensored formula, which works better!!!
* Fix a crash when using weird resolutions. Remove an unnecessary UNet call
* Improve comments, optimize memory in blur routine
* SAG works with sampler_cfg_function
1 year ago
comfyanonymous
b454a67bb9
Support segmind vega model.
1 year ago
comfyanonymous
824e4935f5
Add dtype parameter to VAE object.
1 year ago
comfyanonymous
32b7e7e769
Add manual cast to controlnet.
1 year ago
comfyanonymous
3152023fbc
Use inference dtype for unet memory usage estimation.
1 year ago
comfyanonymous
77755ab8db
Refactor comfy.ops
...
comfy.ops -> comfy.ops.disable_weight_init
This should make it more clear what they actually do.
Some unused code has also been removed.
1 year ago
comfyanonymous
b0aab1e4ea
Add an option --fp16-unet to force using fp16 for the unet.
1 year ago
comfyanonymous
ba07cb748e
Use faster manual cast for fp8 in unet.
1 year ago
comfyanonymous
57926635e8
Switch text encoder to manual cast.
...
Use fp16 text encoder weights for CPU inference to lower memory usage.
1 year ago
comfyanonymous
340177e6e8
Disable non blocking on mps.
1 year ago
comfyanonymous
614b7e731f
Implement GLora.
1 year ago
comfyanonymous
cb63e230b4
Make lora code a bit cleaner.
1 year ago
comfyanonymous
174eba8e95
Use own clip vision model implementation.
1 year ago
comfyanonymous
97015b6b38
Cleanup.
1 year ago
comfyanonymous
a4ec54a40d
Add linear_start and linear_end to model_config.sampling_settings
1 year ago
comfyanonymous
9ac0b487ac
Make --gpu-only put intermediate values in GPU memory instead of cpu.
1 year ago
comfyanonymous
efb704c758
Support attention masking in CLIP implementation.
1 year ago
comfyanonymous
fbdb14d4c4
Cleaner CLIP text encoder implementation.
...
Use a simple CLIP model implementation instead of the one from
transformers.
This will allow some interesting things that would too hackish to implement
using the transformers implementation.
1 year ago
comfyanonymous
2db86b4676
Slightly faster lora applying.
1 year ago
comfyanonymous
1bbd65ab30
Missed this one.
1 year ago
comfyanonymous
9b655d4fd7
Fix memory issue with control loras.
1 year ago
comfyanonymous
26b1c0a771
Fix control lora on fp8.
1 year ago
comfyanonymous
be3468ddd5
Less useless downcasting.
1 year ago
comfyanonymous
ca82ade765
Use .itemsize to get dtype size for fp8.
1 year ago
comfyanonymous
31b0f6f3d8
UNET weights can now be stored in fp8.
...
--fp8_e4m3fn-unet and --fp8_e5m2-unet are the two different formats
supported by pytorch.
1 year ago
comfyanonymous
af365e4dd1
All the unet ops with weights are now handled by comfy.ops
1 year ago
comfyanonymous
61a123a1e0
A different way of handling multiple images passed to SVD.
...
Previously when a list of 3 images [0, 1, 2] was used for a 6 frame video
they were concated like this:
[0, 1, 2, 0, 1, 2]
now they are concated like this:
[0, 0, 1, 1, 2, 2]
1 year ago
comfyanonymous
c97be4db91
Support SD2.1 turbo checkpoint.
1 year ago
comfyanonymous
983ebc5792
Use smart model management for VAE to decrease latency.
1 year ago
comfyanonymous
c45d1b9b67
Add a function to load a unet from a state dict.
1 year ago
comfyanonymous
f30b992b18
.sigma and .timestep now return tensors on the same device as the input.
1 year ago
comfyanonymous
13fdee6abf
Try to free memory for both cond+uncond before inference.
1 year ago
comfyanonymous
be71bb5e13
Tweak memory inference calculations a bit.
1 year ago
comfyanonymous
39e75862b2
Fix regression from last commit.
1 year ago