186 Commits (517f4a94e4a5c45edc64594d70585ec8aeb787e0)

Author SHA1 Message Date
comfyanonymous ad76574cb8 Fix some potential issues with the previous commits. 7 months ago
comfyanonymous 5c69cde037 Load TE model straight to vram if certain conditions are met. 7 months ago
comfyanonymous 1de69fe4d5 Fix some issues with inference slowing down. 7 months ago
comfyanonymous 55ad9d5f8c Fix regression. 7 months ago
comfyanonymous 037c38eb0f Try to improve inference speed on some machines. 7 months ago
comfyanonymous 66d4233210 Fix. 7 months ago
comfyanonymous 08f92d55e9 Partial model shift support. 7 months ago
comfyanonymous 6969fc9ba4 Make supported_dtypes a priority list. 7 months ago
comfyanonymous b334605a66 Fix OOMs happening in some cases.
A cloned model patcher sometimes reported a model was loaded on a device
when it wasn't.
7 months ago
comfyanonymous c14ac98fed Unload models and load them back in lowvram mode no free vram. 7 months ago
comfyanonymous 8edbcf5209 Improve performance on some lowend GPUs. 7 months ago
comfyanonymous f7a5107784 Fix crash. 7 months ago
comfyanonymous 91be9c2867 Tweak lowvram memory formula. 7 months ago
comfyanonymous 03c5018c98 Lower lowvram memory to 1/3 of free memory. 7 months ago
comfyanonymous 2ba5cc8b86 Fix some issues. 7 months ago
comfyanonymous 1e68002b87 Cap lowvram to half of free memory. 7 months ago
comfyanonymous ba9095e5bd Automatically use fp8 for diffusion model weights if:
Checkpoint contains weights in fp8.

There isn't enough memory to load the diffusion model in GPU vram.
7 months ago
comfyanonymous d965474aaa Make ComfyUI split batches a higher priority than weight offload. 7 months ago
comfyanonymous a6decf1e62 Fix bfloat16 potentially not being enabled on mps. 7 months ago
comfyanonymous 1aa9cf3292 Make lowvram more aggressive on low memory machines. 7 months ago
comfyanonymous 5f98de7697 Load flux t5 in fp8 if weights are in fp8. 7 months ago
comfyanonymous 7ad574bffd Mac supports bf16 just make sure you are using the latest pytorch. 7 months ago
comfyanonymous e2382b6adb Make lowvram less aggressive when there are large amounts of free memory. 7 months ago
comfyanonymous 6425252c4f Use fp16 as the default vae dtype for the audio VAE. 8 months ago
comfyanonymous 0ec513d877 Add a --force-channels-last to inference models in channel last mode. 8 months ago
Simon Lui 5eb98f0092
Exempt IPEX from non_blocking previews fixing segmentation faults. (#3708) 8 months ago
comfyanonymous 0e49211a11 Load the SD3 T5xxl model in the same dtype stored in the checkpoint. 9 months ago
comfyanonymous 104fcea0c8 Add function to get the list of currently loaded models. 9 months ago
comfyanonymous b1fd26fe9e pytorch xpu should be flash or mem efficient attention? 9 months ago
comfyanonymous b249862080 Add an annoying print to a function I want to remove. 9 months ago
comfyanonymous bf3e334d46 Disable non_blocking when --deterministic or directml. 9 months ago
comfyanonymous 0920e0e5fe Remove some unused imports. 9 months ago
comfyanonymous 6c23854f54 Fix OSX latent2rgb previews. 9 months ago
comfyanonymous 8508df2569 Work around black image bug on Mac 14.5 by forcing attention upcasting. 9 months ago
comfyanonymous 09e069ae6c Log the pytorch version. 9 months ago
comfyanonymous 19300655dd Don't automatically switch to lowvram mode on GPUs with low memory. 9 months ago
Simon Lui f509c6fe21
Fix Intel GPU memory allocation accuracy and documentation update. (#3459)
* Change calculation of memory total to be more accurate, allocated is actually smaller than reserved.

* Update README.md install documentation for Intel GPUs.
10 months ago
comfyanonymous fa6dd7e5bb Fix lowvram issue with saving checkpoints.
The previous fix didn't cover the case where the model was loaded in
lowvram mode right before.
10 months ago
comfyanonymous 49c20cdc70 No longer necessary. 10 months ago
comfyanonymous e1489ad257 Fix issue with lowvram mode breaking model saving. 10 months ago
Simon Lui a56d02efc7
Change torch.xpu to ipex.optimize, xpu device initialization and remove workaround for text node issue from older IPEX. (#3388) 10 months ago
comfyanonymous 258dbc06c3 Fix some memory related issues. 11 months ago
comfyanonymous 0a03009808 Fix issue with controlnet models getting loaded multiple times. 11 months ago
comfyanonymous 5d8898c056 Fix some performance issues with weight loading and unloading.
Lower peak memory usage when changing model.

Fix case where model weights would be unloaded and reloaded.
11 months ago
comfyanonymous c6de09b02e Optimize memory unload strategy for more optimized performance. 11 months ago
comfyanonymous 4b9005e949 Fix regression with model merging. 11 months ago
comfyanonymous c18a203a8a Don't unload model weights for non weight patches. 11 months ago
comfyanonymous db8b59ecff Lower memory usage for loras in lowvram mode at the cost of perf. 12 months ago
comfyanonymous 0ed72befe1 Change log levels.
Logging level now defaults to info. --verbose sets it to debug.
12 months ago
comfyanonymous 65397ce601 Replace prints with logging and add --verbose argument. 12 months ago