From cd54d502fcc6f47d76b8da781b415fc47c345f74 Mon Sep 17 00:00:00 2001 From: Jedrzej Kosinski Date: Mon, 18 Aug 2025 15:34:53 -0700 Subject: [PATCH] Make sure models_memory_reserve is considered with inference_memory as well in max func calls --- comfy/model_management.py | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/comfy/model_management.py b/comfy/model_management.py index 1f5b9ddb8..2c702959a 100644 --- a/comfy/model_management.py +++ b/comfy/model_management.py @@ -594,11 +594,11 @@ def load_models_gpu(models, memory_required=0, force_patch_weights=False, minimu inference_memory = minimum_inference_memory() models_memory_reserve = get_models_memory_reserve(models) - extra_mem = max(inference_memory, memory_required + extra_reserved_memory() + models_memory_reserve) + extra_mem = max(inference_memory + models_memory_reserve, memory_required + extra_reserved_memory() + models_memory_reserve) if minimum_memory_required is None: minimum_memory_required = extra_mem else: - minimum_memory_required = max(inference_memory, minimum_memory_required + extra_reserved_memory() + models_memory_reserve) + minimum_memory_required = max(inference_memory + models_memory_reserve, minimum_memory_required + extra_reserved_memory() + models_memory_reserve) models = set(models)