feat: Enhance WanVideoLoraSelect with improved low_mem_load and merge_loras options for better LORA management, see #363

This commit is contained in:
Will Miao
2025-08-18 15:05:57 +08:00
parent ca692ed0f2
commit 9c45d9db6c
2 changed files with 12 additions and 7 deletions

View File

@@ -14,7 +14,8 @@ class WanVideoLoraSelect:
def INPUT_TYPES(cls):
return {
"required": {
"low_mem_load": ("BOOLEAN", {"default": False, "tooltip": "Load the LORA model with less VRAM usage, slower loading"}),
"low_mem_load": ("BOOLEAN", {"default": False, "tooltip": "Load LORA models with less VRAM usage, slower loading. This affects ALL LoRAs, not just the current ones. No effect if merge_loras is False"}),
"merge_loras": ("BOOLEAN", {"default": True, "tooltip": "Merge LoRAs into the model, otherwise they are loaded on the fly. Always disabled for GGUF and scaled fp8 models. This affects ALL LoRAs, not just the current one"}),
"text": (IO.STRING, {
"multiline": True,
"dynamicPrompts": True,
@@ -29,7 +30,7 @@ class WanVideoLoraSelect:
RETURN_NAMES = ("lora", "trigger_words", "active_loras")
FUNCTION = "process_loras"
def process_loras(self, text, low_mem_load=False, **kwargs):
def process_loras(self, text, low_mem_load=False, merge_loras=True, **kwargs):
loras_list = []
all_trigger_words = []
active_loras = []
@@ -38,6 +39,9 @@ class WanVideoLoraSelect:
prev_lora = kwargs.get('prev_lora', None)
if prev_lora is not None:
loras_list.extend(prev_lora)
if not merge_loras:
low_mem_load = False # Unmerged LoRAs don't need low_mem_load
# Get blocks if available
blocks = kwargs.get('blocks', {})
@@ -65,6 +69,7 @@ class WanVideoLoraSelect:
"blocks": selected_blocks,
"layer_filter": layer_filter,
"low_mem_load": low_mem_load,
"merge_loras": merge_loras,
}
# Add to list and collect active loras

View File

@@ -30,12 +30,12 @@ app.registerExtension({
// Restore saved value if exists
let existingLoras = [];
if (this.widgets_values && this.widgets_values.length > 0) {
// 0 for low_mem_load, 1 for text widget, 2 for loras widget
const savedValue = this.widgets_values[2];
// 0 for low_mem_load, 1 for merge_loras, 2 for text widget, 3 for loras widget
const savedValue = this.widgets_values[3];
existingLoras = savedValue || [];
}
// Merge the loras data
const mergedLoras = mergeLoras(this.widgets[1].value, existingLoras);
const mergedLoras = mergeLoras(this.widgets[2].value, existingLoras);
// Add flag to prevent callback loops
let isUpdating = false;
@@ -49,7 +49,7 @@ app.registerExtension({
try {
// Remove loras that are not in the value array
const inputWidget = this.widgets[1];
const inputWidget = this.widgets[2];
const currentLoras = value.map(l => l.name);
// Use the constant pattern here as well
@@ -79,7 +79,7 @@ app.registerExtension({
this.lorasWidget = result.widget;
// Update input widget callback
const inputWidget = this.widgets[1];
const inputWidget = this.widgets[2];
inputWidget.options.getMaxHeight = () => 100;
this.inputWidget = inputWidget;
// Wrap the callback with autocomplete setup