diff --git a/comfy/model_management.py b/comfy/model_management.py index 0471932..c54a360 100644 --- a/comfy/model_management.py +++ b/comfy/model_management.py @@ -446,7 +446,7 @@ def load_models_gpu(models, memory_required=0, force_patch_weights=False): if lowvram_available and (vram_set_state == VRAMState.LOW_VRAM or vram_set_state == VRAMState.NORMAL_VRAM): model_size = loaded_model.model_memory_required(torch_dev) current_free_mem = get_free_memory(torch_dev) - lowvram_model_memory = int(max(64 * (1024 * 1024), (current_free_mem - 1024 * (1024 * 1024)) / 1.3 )) + lowvram_model_memory = int(max(64 * (1024 * 1024), (current_free_mem - extra_mem))) if model_size <= (current_free_mem - inference_memory): #only switch to lowvram if really necessary lowvram_model_memory = 0