From b54bf359bf6145ff7a1a2f676079ebac339c862b Mon Sep 17 00:00:00 2001 From: Manuel Schmid Date: Mon, 3 Feb 2025 04:11:49 +0100 Subject: [PATCH] sd_api_pictures model reload fix (#6720) --- extensions/sd_api_pictures/script.py | 9 ++++++--- 1 file changed, 6 insertions(+), 3 deletions(-) diff --git a/extensions/sd_api_pictures/script.py b/extensions/sd_api_pictures/script.py index 3a31771af6..f216da3849 100644 --- a/extensions/sd_api_pictures/script.py +++ b/extensions/sd_api_pictures/script.py @@ -11,7 +11,7 @@ from PIL import Image from modules import shared -from modules.models import reload_model, unload_model +from modules.models import load_model, unload_model from modules.ui import create_refresh_button torch._C._jit_set_profiling_mode(False) @@ -38,7 +38,8 @@ 'cfg_scale': 7, 'textgen_prefix': 'Please provide a detailed and vivid description of [subject]', 'sd_checkpoint': ' ', - 'checkpoint_list': [" "] + 'checkpoint_list': [" "], + 'last_model': "" } @@ -46,6 +47,7 @@ def give_VRAM_priority(actor): global shared, params if actor == 'SD': + params["last_model"] = shared.model_name unload_model() print("Requesting Auto1111 to re-load last checkpoint used...") response = requests.post(url=f'{params["address"]}/sdapi/v1/reload-checkpoint', json='') @@ -55,7 +57,8 @@ def give_VRAM_priority(actor): print("Requesting Auto1111 to vacate VRAM...") response = requests.post(url=f'{params["address"]}/sdapi/v1/unload-checkpoint', json='') response.raise_for_status() - reload_model() + if params["last_model"]: + shared.model, shared.tokenizer = load_model(params["last_model"]) elif actor == 'set': print("VRAM mangement activated -- requesting Auto1111 to vacate VRAM...")