mirror of
https://github.com/willmiao/ComfyUI-Lora-Manager.git
synced 2026-03-25 07:05:43 -03:00
30
py/config.py
30
py/config.py
@@ -103,21 +103,29 @@ class Config:
|
|||||||
|
|
||||||
def _init_lora_paths(self) -> List[str]:
|
def _init_lora_paths(self) -> List[str]:
|
||||||
"""Initialize and validate LoRA paths from ComfyUI settings"""
|
"""Initialize and validate LoRA paths from ComfyUI settings"""
|
||||||
paths = sorted(set(path.replace(os.sep, "/")
|
raw_paths = folder_paths.get_folder_paths("loras")
|
||||||
for path in folder_paths.get_folder_paths("loras")
|
|
||||||
if os.path.exists(path)), key=lambda p: p.lower())
|
|
||||||
print("Found LoRA roots:", "\n - " + "\n - ".join(paths))
|
|
||||||
|
|
||||||
if not paths:
|
# Normalize and resolve symlinks, store mapping from resolved -> original
|
||||||
|
path_map = {}
|
||||||
|
for path in raw_paths:
|
||||||
|
if os.path.exists(path):
|
||||||
|
real_path = os.path.normpath(os.path.realpath(path)).replace(os.sep, '/')
|
||||||
|
path_map[real_path] = path_map.get(real_path, path) # preserve first seen
|
||||||
|
|
||||||
|
# Now sort and use only the deduplicated real paths
|
||||||
|
unique_paths = sorted(path_map.values(), key=lambda p: p.lower())
|
||||||
|
print("Found LoRA roots:", "\n - " + "\n - ".join(unique_paths))
|
||||||
|
|
||||||
|
if not unique_paths:
|
||||||
raise ValueError("No valid loras folders found in ComfyUI configuration")
|
raise ValueError("No valid loras folders found in ComfyUI configuration")
|
||||||
|
|
||||||
# 初始化路径映射
|
for original_path in unique_paths:
|
||||||
for path in paths:
|
real_path = os.path.normpath(os.path.realpath(original_path)).replace(os.sep, '/')
|
||||||
real_path = os.path.normpath(os.path.realpath(path)).replace(os.sep, '/')
|
if real_path != original_path:
|
||||||
if real_path != path:
|
self.add_path_mapping(original_path, real_path)
|
||||||
self.add_path_mapping(path, real_path)
|
|
||||||
|
|
||||||
return paths
|
return unique_paths
|
||||||
|
|
||||||
|
|
||||||
def _init_checkpoint_paths(self) -> List[str]:
|
def _init_checkpoint_paths(self) -> List[str]:
|
||||||
"""Initialize and validate checkpoint paths from ComfyUI settings"""
|
"""Initialize and validate checkpoint paths from ComfyUI settings"""
|
||||||
|
|||||||
@@ -5,7 +5,7 @@ from ..services.lora_scanner import LoraScanner
|
|||||||
from ..config import config
|
from ..config import config
|
||||||
import asyncio
|
import asyncio
|
||||||
import os
|
import os
|
||||||
from .utils import FlexibleOptionalInputType, any_type
|
from .utils import FlexibleOptionalInputType, any_type, get_lora_info, extract_lora_name, get_loras_list
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
@@ -32,48 +32,6 @@ class LoraManagerLoader:
|
|||||||
RETURN_TYPES = ("MODEL", "CLIP", IO.STRING, IO.STRING)
|
RETURN_TYPES = ("MODEL", "CLIP", IO.STRING, IO.STRING)
|
||||||
RETURN_NAMES = ("MODEL", "CLIP", "trigger_words", "loaded_loras")
|
RETURN_NAMES = ("MODEL", "CLIP", "trigger_words", "loaded_loras")
|
||||||
FUNCTION = "load_loras"
|
FUNCTION = "load_loras"
|
||||||
|
|
||||||
async def get_lora_info(self, lora_name):
|
|
||||||
"""Get the lora path and trigger words from cache"""
|
|
||||||
scanner = await LoraScanner.get_instance()
|
|
||||||
cache = await scanner.get_cached_data()
|
|
||||||
|
|
||||||
for item in cache.raw_data:
|
|
||||||
if item.get('file_name') == lora_name:
|
|
||||||
file_path = item.get('file_path')
|
|
||||||
if file_path:
|
|
||||||
for root in config.loras_roots:
|
|
||||||
root = root.replace(os.sep, '/')
|
|
||||||
if file_path.startswith(root):
|
|
||||||
relative_path = os.path.relpath(file_path, root).replace(os.sep, '/')
|
|
||||||
# Get trigger words from civitai metadata
|
|
||||||
civitai = item.get('civitai', {})
|
|
||||||
trigger_words = civitai.get('trainedWords', []) if civitai else []
|
|
||||||
return relative_path, trigger_words
|
|
||||||
return lora_name, [] # Fallback if not found
|
|
||||||
|
|
||||||
def extract_lora_name(self, lora_path):
|
|
||||||
"""Extract the lora name from a lora path (e.g., 'IL\\aorunIllstrious.safetensors' -> 'aorunIllstrious')"""
|
|
||||||
# Get the basename without extension
|
|
||||||
basename = os.path.basename(lora_path)
|
|
||||||
return os.path.splitext(basename)[0]
|
|
||||||
|
|
||||||
def _get_loras_list(self, kwargs):
|
|
||||||
"""Helper to extract loras list from either old or new kwargs format"""
|
|
||||||
if 'loras' not in kwargs:
|
|
||||||
return []
|
|
||||||
|
|
||||||
loras_data = kwargs['loras']
|
|
||||||
# Handle new format: {'loras': {'__value__': [...]}}
|
|
||||||
if isinstance(loras_data, dict) and '__value__' in loras_data:
|
|
||||||
return loras_data['__value__']
|
|
||||||
# Handle old format: {'loras': [...]}
|
|
||||||
elif isinstance(loras_data, list):
|
|
||||||
return loras_data
|
|
||||||
# Unexpected format
|
|
||||||
else:
|
|
||||||
logger.warning(f"Unexpected loras format: {type(loras_data)}")
|
|
||||||
return []
|
|
||||||
|
|
||||||
def load_loras(self, model, text, **kwargs):
|
def load_loras(self, model, text, **kwargs):
|
||||||
"""Loads multiple LoRAs based on the kwargs input and lora_stack."""
|
"""Loads multiple LoRAs based on the kwargs input and lora_stack."""
|
||||||
@@ -89,14 +47,14 @@ class LoraManagerLoader:
|
|||||||
model, clip = LoraLoader().load_lora(model, clip, lora_path, model_strength, clip_strength)
|
model, clip = LoraLoader().load_lora(model, clip, lora_path, model_strength, clip_strength)
|
||||||
|
|
||||||
# Extract lora name for trigger words lookup
|
# Extract lora name for trigger words lookup
|
||||||
lora_name = self.extract_lora_name(lora_path)
|
lora_name = extract_lora_name(lora_path)
|
||||||
_, trigger_words = asyncio.run(self.get_lora_info(lora_name))
|
_, trigger_words = asyncio.run(get_lora_info(lora_name))
|
||||||
|
|
||||||
all_trigger_words.extend(trigger_words)
|
all_trigger_words.extend(trigger_words)
|
||||||
loaded_loras.append(f"{lora_name}: {model_strength}")
|
loaded_loras.append(f"{lora_name}: {model_strength}")
|
||||||
|
|
||||||
# Then process loras from kwargs with support for both old and new formats
|
# Then process loras from kwargs with support for both old and new formats
|
||||||
loras_list = self._get_loras_list(kwargs)
|
loras_list = get_loras_list(kwargs)
|
||||||
for lora in loras_list:
|
for lora in loras_list:
|
||||||
if not lora.get('active', False):
|
if not lora.get('active', False):
|
||||||
continue
|
continue
|
||||||
@@ -105,7 +63,7 @@ class LoraManagerLoader:
|
|||||||
strength = float(lora['strength'])
|
strength = float(lora['strength'])
|
||||||
|
|
||||||
# Get lora path and trigger words
|
# Get lora path and trigger words
|
||||||
lora_path, trigger_words = asyncio.run(self.get_lora_info(lora_name))
|
lora_path, trigger_words = asyncio.run(get_lora_info(lora_name))
|
||||||
|
|
||||||
# Apply the LoRA using the resolved path
|
# Apply the LoRA using the resolved path
|
||||||
model, clip = LoraLoader().load_lora(model, clip, lora_path, strength, strength)
|
model, clip = LoraLoader().load_lora(model, clip, lora_path, strength, strength)
|
||||||
|
|||||||
@@ -3,7 +3,7 @@ from ..services.lora_scanner import LoraScanner
|
|||||||
from ..config import config
|
from ..config import config
|
||||||
import asyncio
|
import asyncio
|
||||||
import os
|
import os
|
||||||
from .utils import FlexibleOptionalInputType, any_type
|
from .utils import FlexibleOptionalInputType, any_type, get_lora_info, extract_lora_name, get_loras_list
|
||||||
import logging
|
import logging
|
||||||
|
|
||||||
logger = logging.getLogger(__name__)
|
logger = logging.getLogger(__name__)
|
||||||
@@ -29,48 +29,6 @@ class LoraStacker:
|
|||||||
RETURN_TYPES = ("LORA_STACK", IO.STRING, IO.STRING)
|
RETURN_TYPES = ("LORA_STACK", IO.STRING, IO.STRING)
|
||||||
RETURN_NAMES = ("LORA_STACK", "trigger_words", "active_loras")
|
RETURN_NAMES = ("LORA_STACK", "trigger_words", "active_loras")
|
||||||
FUNCTION = "stack_loras"
|
FUNCTION = "stack_loras"
|
||||||
|
|
||||||
async def get_lora_info(self, lora_name):
|
|
||||||
"""Get the lora path and trigger words from cache"""
|
|
||||||
scanner = await LoraScanner.get_instance()
|
|
||||||
cache = await scanner.get_cached_data()
|
|
||||||
|
|
||||||
for item in cache.raw_data:
|
|
||||||
if item.get('file_name') == lora_name:
|
|
||||||
file_path = item.get('file_path')
|
|
||||||
if file_path:
|
|
||||||
for root in config.loras_roots:
|
|
||||||
root = root.replace(os.sep, '/')
|
|
||||||
if file_path.startswith(root):
|
|
||||||
relative_path = os.path.relpath(file_path, root).replace(os.sep, '/')
|
|
||||||
# Get trigger words from civitai metadata
|
|
||||||
civitai = item.get('civitai', {})
|
|
||||||
trigger_words = civitai.get('trainedWords', []) if civitai else []
|
|
||||||
return relative_path, trigger_words
|
|
||||||
return lora_name, [] # Fallback if not found
|
|
||||||
|
|
||||||
def extract_lora_name(self, lora_path):
|
|
||||||
"""Extract the lora name from a lora path (e.g., 'IL\\aorunIllstrious.safetensors' -> 'aorunIllstrious')"""
|
|
||||||
# Get the basename without extension
|
|
||||||
basename = os.path.basename(lora_path)
|
|
||||||
return os.path.splitext(basename)[0]
|
|
||||||
|
|
||||||
def _get_loras_list(self, kwargs):
|
|
||||||
"""Helper to extract loras list from either old or new kwargs format"""
|
|
||||||
if 'loras' not in kwargs:
|
|
||||||
return []
|
|
||||||
|
|
||||||
loras_data = kwargs['loras']
|
|
||||||
# Handle new format: {'loras': {'__value__': [...]}}
|
|
||||||
if isinstance(loras_data, dict) and '__value__' in loras_data:
|
|
||||||
return loras_data['__value__']
|
|
||||||
# Handle old format: {'loras': [...]}
|
|
||||||
elif isinstance(loras_data, list):
|
|
||||||
return loras_data
|
|
||||||
# Unexpected format
|
|
||||||
else:
|
|
||||||
logger.warning(f"Unexpected loras format: {type(loras_data)}")
|
|
||||||
return []
|
|
||||||
|
|
||||||
def stack_loras(self, text, **kwargs):
|
def stack_loras(self, text, **kwargs):
|
||||||
"""Stacks multiple LoRAs based on the kwargs input without loading them."""
|
"""Stacks multiple LoRAs based on the kwargs input without loading them."""
|
||||||
@@ -84,12 +42,12 @@ class LoraStacker:
|
|||||||
stack.extend(lora_stack)
|
stack.extend(lora_stack)
|
||||||
# Get trigger words from existing stack entries
|
# Get trigger words from existing stack entries
|
||||||
for lora_path, _, _ in lora_stack:
|
for lora_path, _, _ in lora_stack:
|
||||||
lora_name = self.extract_lora_name(lora_path)
|
lora_name = extract_lora_name(lora_path)
|
||||||
_, trigger_words = asyncio.run(self.get_lora_info(lora_name))
|
_, trigger_words = asyncio.run(get_lora_info(lora_name))
|
||||||
all_trigger_words.extend(trigger_words)
|
all_trigger_words.extend(trigger_words)
|
||||||
|
|
||||||
# Process loras from kwargs with support for both old and new formats
|
# Process loras from kwargs with support for both old and new formats
|
||||||
loras_list = self._get_loras_list(kwargs)
|
loras_list = get_loras_list(kwargs)
|
||||||
for lora in loras_list:
|
for lora in loras_list:
|
||||||
if not lora.get('active', False):
|
if not lora.get('active', False):
|
||||||
continue
|
continue
|
||||||
@@ -99,7 +57,7 @@ class LoraStacker:
|
|||||||
clip_strength = model_strength # Using same strength for both as in the original loader
|
clip_strength = model_strength # Using same strength for both as in the original loader
|
||||||
|
|
||||||
# Get lora path and trigger words
|
# Get lora path and trigger words
|
||||||
lora_path, trigger_words = asyncio.run(self.get_lora_info(lora_name))
|
lora_path, trigger_words = asyncio.run(get_lora_info(lora_name))
|
||||||
|
|
||||||
# Add to stack without loading
|
# Add to stack without loading
|
||||||
# replace '/' with os.sep to avoid different OS path format
|
# replace '/' with os.sep to avoid different OS path format
|
||||||
|
|||||||
@@ -47,10 +47,10 @@ class TriggerWordToggle:
|
|||||||
trigger_words = trigger_words_data if isinstance(trigger_words_data, str) else ""
|
trigger_words = trigger_words_data if isinstance(trigger_words_data, str) else ""
|
||||||
|
|
||||||
# Send trigger words to frontend
|
# Send trigger words to frontend
|
||||||
PromptServer.instance.send_sync("trigger_word_update", {
|
# PromptServer.instance.send_sync("trigger_word_update", {
|
||||||
"id": id,
|
# "id": id,
|
||||||
"message": trigger_words
|
# "message": trigger_words
|
||||||
})
|
# })
|
||||||
|
|
||||||
filtered_triggers = trigger_words
|
filtered_triggers = trigger_words
|
||||||
|
|
||||||
|
|||||||
@@ -30,4 +30,55 @@ class FlexibleOptionalInputType(dict):
|
|||||||
return True
|
return True
|
||||||
|
|
||||||
|
|
||||||
any_type = AnyType("*")
|
any_type = AnyType("*")
|
||||||
|
|
||||||
|
# Common methods extracted from lora_loader.py and lora_stacker.py
|
||||||
|
import os
|
||||||
|
import logging
|
||||||
|
import asyncio
|
||||||
|
from ..services.lora_scanner import LoraScanner
|
||||||
|
from ..config import config
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
async def get_lora_info(lora_name):
|
||||||
|
"""Get the lora path and trigger words from cache"""
|
||||||
|
scanner = await LoraScanner.get_instance()
|
||||||
|
cache = await scanner.get_cached_data()
|
||||||
|
|
||||||
|
for item in cache.raw_data:
|
||||||
|
if item.get('file_name') == lora_name:
|
||||||
|
file_path = item.get('file_path')
|
||||||
|
if file_path:
|
||||||
|
for root in config.loras_roots:
|
||||||
|
root = root.replace(os.sep, '/')
|
||||||
|
if file_path.startswith(root):
|
||||||
|
relative_path = os.path.relpath(file_path, root).replace(os.sep, '/')
|
||||||
|
# Get trigger words from civitai metadata
|
||||||
|
civitai = item.get('civitai', {})
|
||||||
|
trigger_words = civitai.get('trainedWords', []) if civitai else []
|
||||||
|
return relative_path, trigger_words
|
||||||
|
return lora_name, [] # Fallback if not found
|
||||||
|
|
||||||
|
def extract_lora_name(lora_path):
|
||||||
|
"""Extract the lora name from a lora path (e.g., 'IL\\aorunIllstrious.safetensors' -> 'aorunIllstrious')"""
|
||||||
|
# Get the basename without extension
|
||||||
|
basename = os.path.basename(lora_path)
|
||||||
|
return os.path.splitext(basename)[0]
|
||||||
|
|
||||||
|
def get_loras_list(kwargs):
|
||||||
|
"""Helper to extract loras list from either old or new kwargs format"""
|
||||||
|
if 'loras' not in kwargs:
|
||||||
|
return []
|
||||||
|
|
||||||
|
loras_data = kwargs['loras']
|
||||||
|
# Handle new format: {'loras': {'__value__': [...]}}
|
||||||
|
if isinstance(loras_data, dict) and '__value__' in loras_data:
|
||||||
|
return loras_data['__value__']
|
||||||
|
# Handle old format: {'loras': [...]}
|
||||||
|
elif isinstance(loras_data, list):
|
||||||
|
return loras_data
|
||||||
|
# Unexpected format
|
||||||
|
else:
|
||||||
|
logger.warning(f"Unexpected loras format: {type(loras_data)}")
|
||||||
|
return []
|
||||||
@@ -3,8 +3,10 @@ import json
|
|||||||
import logging
|
import logging
|
||||||
from aiohttp import web
|
from aiohttp import web
|
||||||
from typing import Dict
|
from typing import Dict
|
||||||
|
from server import PromptServer # type: ignore
|
||||||
|
|
||||||
from ..utils.routes_common import ModelRouteUtils
|
from ..utils.routes_common import ModelRouteUtils
|
||||||
|
from ..nodes.utils import get_lora_info
|
||||||
|
|
||||||
from ..config import config
|
from ..config import config
|
||||||
from ..services.websocket_manager import ws_manager
|
from ..services.websocket_manager import ws_manager
|
||||||
@@ -64,6 +66,9 @@ class ApiRoutes:
|
|||||||
app.router.add_get('/api/lora-civitai-url', routes.get_lora_civitai_url) # Add new route for Civitai URL
|
app.router.add_get('/api/lora-civitai-url', routes.get_lora_civitai_url) # Add new route for Civitai URL
|
||||||
app.router.add_post('/api/rename_lora', routes.rename_lora) # Add new route for renaming LoRA files
|
app.router.add_post('/api/rename_lora', routes.rename_lora) # Add new route for renaming LoRA files
|
||||||
app.router.add_get('/api/loras/scan', routes.scan_loras) # Add new route for scanning LoRA files
|
app.router.add_get('/api/loras/scan', routes.scan_loras) # Add new route for scanning LoRA files
|
||||||
|
|
||||||
|
# Add the new trigger words route
|
||||||
|
app.router.add_post('/loramanager/get_trigger_words', routes.get_trigger_words)
|
||||||
|
|
||||||
# Add update check routes
|
# Add update check routes
|
||||||
UpdateRoutes.setup_routes(app)
|
UpdateRoutes.setup_routes(app)
|
||||||
@@ -1021,4 +1026,35 @@ class ApiRoutes:
|
|||||||
return web.json_response({
|
return web.json_response({
|
||||||
'success': False,
|
'success': False,
|
||||||
'error': str(e)
|
'error': str(e)
|
||||||
|
}, status=500)
|
||||||
|
|
||||||
|
async def get_trigger_words(self, request: web.Request) -> web.Response:
|
||||||
|
"""Get trigger words for specified LoRA models"""
|
||||||
|
try:
|
||||||
|
json_data = await request.json()
|
||||||
|
lora_names = json_data.get("lora_names", [])
|
||||||
|
node_ids = json_data.get("node_ids", [])
|
||||||
|
|
||||||
|
all_trigger_words = []
|
||||||
|
for lora_name in lora_names:
|
||||||
|
_, trigger_words = await get_lora_info(lora_name)
|
||||||
|
all_trigger_words.extend(trigger_words)
|
||||||
|
|
||||||
|
# Format the trigger words
|
||||||
|
trigger_words_text = ",, ".join(all_trigger_words) if all_trigger_words else ""
|
||||||
|
|
||||||
|
# Send update to all connected trigger word toggle nodes
|
||||||
|
for node_id in node_ids:
|
||||||
|
PromptServer.instance.send_sync("trigger_word_update", {
|
||||||
|
"id": node_id,
|
||||||
|
"message": trigger_words_text
|
||||||
|
})
|
||||||
|
|
||||||
|
return web.json_response({"success": True})
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
logger.error(f"Error getting trigger words: {e}")
|
||||||
|
return web.json_response({
|
||||||
|
"success": False,
|
||||||
|
"error": str(e)
|
||||||
}, status=500)
|
}, status=500)
|
||||||
26
py/server_routes.py
Normal file
26
py/server_routes.py
Normal file
@@ -0,0 +1,26 @@
|
|||||||
|
from aiohttp import web
|
||||||
|
from server import PromptServer
|
||||||
|
from .nodes.utils import get_lora_info
|
||||||
|
|
||||||
|
@PromptServer.instance.routes.post("/loramanager/get_trigger_words")
|
||||||
|
async def get_trigger_words(request):
|
||||||
|
json_data = await request.json()
|
||||||
|
lora_names = json_data.get("lora_names", [])
|
||||||
|
node_ids = json_data.get("node_ids", [])
|
||||||
|
|
||||||
|
all_trigger_words = []
|
||||||
|
for lora_name in lora_names:
|
||||||
|
_, trigger_words = await get_lora_info(lora_name)
|
||||||
|
all_trigger_words.extend(trigger_words)
|
||||||
|
|
||||||
|
# Format the trigger words
|
||||||
|
trigger_words_text = ",, ".join(all_trigger_words) if all_trigger_words else ""
|
||||||
|
|
||||||
|
# Send update to all connected trigger word toggle nodes
|
||||||
|
for node_id in node_ids:
|
||||||
|
PromptServer.instance.send_sync("trigger_word_update", {
|
||||||
|
"id": node_id,
|
||||||
|
"message": trigger_words_text
|
||||||
|
})
|
||||||
|
|
||||||
|
return web.json_response({"success": True})
|
||||||
@@ -9,6 +9,57 @@ async function getLorasWidgetModule() {
|
|||||||
return await dynamicImportByVersion("./loras_widget.js", "./legacy_loras_widget.js");
|
return await dynamicImportByVersion("./loras_widget.js", "./legacy_loras_widget.js");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// Function to get connected trigger toggle nodes
|
||||||
|
function getConnectedTriggerToggleNodes(node) {
|
||||||
|
const connectedNodes = [];
|
||||||
|
|
||||||
|
// Check if node has outputs
|
||||||
|
if (node.outputs && node.outputs.length > 0) {
|
||||||
|
// For each output slot
|
||||||
|
for (const output of node.outputs) {
|
||||||
|
// Check if this output has any links
|
||||||
|
if (output.links && output.links.length > 0) {
|
||||||
|
// For each link, get the target node
|
||||||
|
for (const linkId of output.links) {
|
||||||
|
const link = app.graph.links[linkId];
|
||||||
|
if (link) {
|
||||||
|
const targetNode = app.graph.getNodeById(link.target_id);
|
||||||
|
if (targetNode && targetNode.comfyClass === "TriggerWord Toggle (LoraManager)") {
|
||||||
|
connectedNodes.push(targetNode.id);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
return connectedNodes;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Function to update trigger words for connected toggle nodes
|
||||||
|
function updateConnectedTriggerWords(node, text) {
|
||||||
|
const connectedNodeIds = getConnectedTriggerToggleNodes(node);
|
||||||
|
if (connectedNodeIds.length > 0) {
|
||||||
|
// Extract lora names from the text
|
||||||
|
const loraNames = [];
|
||||||
|
let match;
|
||||||
|
// Reset the RegExp object's lastIndex to start from the beginning
|
||||||
|
LORA_PATTERN.lastIndex = 0;
|
||||||
|
while ((match = LORA_PATTERN.exec(text)) !== null) {
|
||||||
|
loraNames.push(match[1]); // match[1] contains the lora name
|
||||||
|
}
|
||||||
|
|
||||||
|
// Call API to get trigger words
|
||||||
|
fetch("/loramanager/get_trigger_words", {
|
||||||
|
method: "POST",
|
||||||
|
headers: { "Content-Type": "application/json" },
|
||||||
|
body: JSON.stringify({
|
||||||
|
lora_names: loraNames,
|
||||||
|
node_ids: connectedNodeIds
|
||||||
|
})
|
||||||
|
}).catch(err => console.error("Error fetching trigger words:", err));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
function mergeLoras(lorasText, lorasArr) {
|
function mergeLoras(lorasText, lorasArr) {
|
||||||
const result = [];
|
const result = [];
|
||||||
let match;
|
let match;
|
||||||
@@ -99,6 +150,9 @@ app.registerExtension({
|
|||||||
newText = newText.replace(/\s+/g, ' ').trim();
|
newText = newText.replace(/\s+/g, ' ').trim();
|
||||||
|
|
||||||
inputWidget.value = newText;
|
inputWidget.value = newText;
|
||||||
|
|
||||||
|
// Add this line to update trigger words when lorasWidget changes cause inputWidget value to change
|
||||||
|
updateConnectedTriggerWords(node, newText);
|
||||||
} finally {
|
} finally {
|
||||||
isUpdating = false;
|
isUpdating = false;
|
||||||
}
|
}
|
||||||
@@ -117,6 +171,9 @@ app.registerExtension({
|
|||||||
const mergedLoras = mergeLoras(value, currentLoras);
|
const mergedLoras = mergeLoras(value, currentLoras);
|
||||||
|
|
||||||
node.lorasWidget.value = mergedLoras;
|
node.lorasWidget.value = mergedLoras;
|
||||||
|
|
||||||
|
// Replace the existing trigger word update code with the new function
|
||||||
|
updateConnectedTriggerWords(node, value);
|
||||||
} finally {
|
} finally {
|
||||||
isUpdating = false;
|
isUpdating = false;
|
||||||
}
|
}
|
||||||
|
|||||||
Reference in New Issue
Block a user