diff --git a/py/nodes/lora_loader.py b/py/nodes/lora_loader.py index ad2b3aa1..2183c8ed 100644 --- a/py/nodes/lora_loader.py +++ b/py/nodes/lora_loader.py @@ -17,7 +17,7 @@ class LoraManagerLoader: "model": ("MODEL",), # "clip": ("CLIP",), "text": ("AUTOCOMPLETE_TEXT_LORAS", { - "placeholder": "Type LoRA syntax...", + "placeholder": "Search LoRAs to add...", "tooltip": "Format: separated by spaces or punctuation", }), }, diff --git a/py/nodes/lora_stacker.py b/py/nodes/lora_stacker.py index 5b67ea83..3a6871f9 100644 --- a/py/nodes/lora_stacker.py +++ b/py/nodes/lora_stacker.py @@ -15,7 +15,7 @@ class LoraStacker: return { "required": { "text": ("AUTOCOMPLETE_TEXT_LORAS", { - "placeholder": "Type LoRA syntax...", + "placeholder": "Search LoRAs to add...", "tooltip": "Format: separated by spaces or punctuation", }), }, diff --git a/py/nodes/wanvideo_lora_select.py b/py/nodes/wanvideo_lora_select.py index 29f8edd3..64b17088 100644 --- a/py/nodes/wanvideo_lora_select.py +++ b/py/nodes/wanvideo_lora_select.py @@ -16,7 +16,7 @@ class WanVideoLoraSelectLM: "low_mem_load": ("BOOLEAN", {"default": False, "tooltip": "Load LORA models with less VRAM usage, slower loading. This affects ALL LoRAs, not just the current ones. No effect if merge_loras is False"}), "merge_loras": ("BOOLEAN", {"default": True, "tooltip": "Merge LoRAs into the model, otherwise they are loaded on the fly. Always disabled for GGUF and scaled fp8 models. This affects ALL LoRAs, not just the current one"}), "text": ("AUTOCOMPLETE_TEXT_LORAS", { - "placeholder": "Type LoRA syntax...", + "placeholder": "Search LoRAs to add...", "tooltip": "Format: separated by spaces or punctuation", }), },