Skip to content

Commit 91be9c2

Browse files
Tweak lowvram memory formula.
1 parent 03c5018 commit 91be9c2

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

comfy/model_management.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -450,7 +450,7 @@ def load_models_gpu(models, memory_required=0, force_patch_weights=False, minimu
450450
if lowvram_available and (vram_set_state == VRAMState.LOW_VRAM or vram_set_state == VRAMState.NORMAL_VRAM):
451451
model_size = loaded_model.model_memory_required(torch_dev)
452452
current_free_mem = get_free_memory(torch_dev)
453-
lowvram_model_memory = max(64 * (1024 * 1024), (current_free_mem - minimum_memory_required), current_free_mem * 0.33)
453+
lowvram_model_memory = max(64 * (1024 * 1024), (current_free_mem - minimum_memory_required), min(current_free_mem * 0.4, current_free_mem - minimum_inference_memory()))
454454
if model_size <= lowvram_model_memory: #only switch to lowvram if really necessary
455455
lowvram_model_memory = 0
456456

0 commit comments

Comments
 (0)