mirror of
https://github.com/idrirap/ComfyUI-Lora-Auto-Trigger-Words.git
synced 2026-03-21 13:12:13 -03:00
166 lines
5.8 KiB
Python
166 lines
5.8 KiB
Python
import folder_paths
|
|
import hashlib
|
|
import json
|
|
import os
|
|
import requests
|
|
|
|
def load_json_from_file(file_path):
|
|
try:
|
|
with open(file_path, 'r') as json_file:
|
|
data = json.load(json_file)
|
|
return data
|
|
except FileNotFoundError:
|
|
print(f"File not found: {file_path}")
|
|
return None
|
|
except json.JSONDecodeError:
|
|
print(f"Error decoding JSON in file: {file_path}")
|
|
raise
|
|
|
|
def save_dict_to_json(data_dict, file_path):
|
|
try:
|
|
with open(file_path, 'w') as json_file:
|
|
json.dump(data_dict, json_file, indent=4)
|
|
print(f"Data saved to {file_path}")
|
|
except Exception as e:
|
|
print(f"Error saving JSON to file: {e}")
|
|
|
|
def get_model_version_info(hash_value):
|
|
api_url = f"https://civitai.com/api/v1/model-versions/by-hash/{hash_value}"
|
|
try:
|
|
response = requests.get(api_url)
|
|
except Exception as e:
|
|
print(f"[Lora-Auto-Trigger] {e}")
|
|
return None
|
|
if response.status_code == 200:
|
|
return response.json()
|
|
else:
|
|
return None
|
|
|
|
def calculate_sha256(file_path):
|
|
sha256_hash = hashlib.sha256()
|
|
with open(file_path, "rb") as f:
|
|
for chunk in iter(lambda: f.read(4096), b""):
|
|
sha256_hash.update(chunk)
|
|
return sha256_hash.hexdigest()
|
|
|
|
|
|
def load_and_save_tags(lora_name, force_fetch):
|
|
json_tags_path = "./loras_tags.json"
|
|
lora_tags = load_json_from_file(json_tags_path)
|
|
output_tags = lora_tags.get(lora_name, None) if lora_tags is not None else None
|
|
if output_tags is not None:
|
|
output_tags_list = output_tags
|
|
else:
|
|
output_tags_list = []
|
|
|
|
lora_path = folder_paths.get_full_path("loras", lora_name)
|
|
if lora_tags is None or force_fetch or output_tags is None: # search on civitai only if no local cache or forced
|
|
print("[Lora-Auto-Trigger] calculating lora hash")
|
|
LORAsha256 = calculate_sha256(lora_path)
|
|
print("[Lora-Auto-Trigger] requesting infos")
|
|
model_info = get_model_version_info(LORAsha256)
|
|
if model_info is not None:
|
|
if "trainedWords" in model_info:
|
|
print("[Lora-Auto-Trigger] tags found!")
|
|
if lora_tags is None:
|
|
lora_tags = {}
|
|
lora_tags[lora_name] = model_info["trainedWords"]
|
|
save_dict_to_json(lora_tags, json_tags_path)
|
|
output_tags_list = model_info["trainedWords"]
|
|
else:
|
|
print("[Lora-Auto-Trigger] No informations found.")
|
|
if lora_tags is None:
|
|
lora_tags = {}
|
|
lora_tags[lora_name] = []
|
|
save_dict_to_json(lora_tags,json_tags_path)
|
|
|
|
return output_tags_list
|
|
|
|
def get_metadata(filepath, type):
|
|
filepath = folder_paths.get_full_path(type, filepath)
|
|
with open(filepath, "rb") as file:
|
|
# https://github.com/huggingface/safetensors#format
|
|
# 8 bytes: N, an unsigned little-endian 64-bit integer, containing the size of the header
|
|
header_size = int.from_bytes(file.read(8), "little", signed=False)
|
|
|
|
if header_size <= 0:
|
|
raise BufferError("Invalid header size")
|
|
|
|
header = file.read(header_size)
|
|
if header_size <= 0:
|
|
raise BufferError("Invalid header")
|
|
header_json = json.loads(header)
|
|
return header_json["__metadata__"] if "__metadata__" in header_json else None
|
|
|
|
# parse the __metadata__ json looking for trained tags
|
|
def sort_tags_by_frequency(meta_tags):
|
|
if meta_tags is None:
|
|
return []
|
|
if "ss_tag_frequency" in meta_tags:
|
|
meta_tags = meta_tags["ss_tag_frequency"]
|
|
meta_tags = json.loads(meta_tags)
|
|
sorted_tags = {}
|
|
for _, dataset in meta_tags.items():
|
|
for tag, count in dataset.items():
|
|
tag = str(tag).strip()
|
|
if tag in sorted_tags:
|
|
sorted_tags[tag] = sorted_tags[tag] + count
|
|
else:
|
|
sorted_tags[tag] = count
|
|
# sort tags by training frequency. Most seen tags firsts
|
|
sorted_tags = dict(sorted(sorted_tags.items(), key=lambda item: item[1], reverse=True))
|
|
return list(sorted_tags.keys())
|
|
else:
|
|
return []
|
|
|
|
def parse_selector(selector, tags_list):
|
|
if len(tags_list) == 0:
|
|
return ""
|
|
range_index_list = selector.split(",")
|
|
output = {}
|
|
for range_index in range_index_list:
|
|
# single value
|
|
if range_index.count(":") == 0:
|
|
# remove empty values
|
|
if range_index.strip() == "":
|
|
continue
|
|
index = int(range_index)
|
|
# ignore out of bound indexes
|
|
if abs(index) > len(tags_list) - 1:
|
|
continue
|
|
output[index] = tags_list[index]
|
|
|
|
# actual range
|
|
if range_index.count(":") == 1:
|
|
indexes = range_index.split(":")
|
|
# check empty
|
|
if indexes[0] == "":
|
|
start = 0
|
|
else:
|
|
start = int(indexes[0])
|
|
if indexes[1] == "":
|
|
end = len(tags_list)
|
|
else:
|
|
end = int(indexes[1])
|
|
# check negative
|
|
if start < 0:
|
|
start = len(tags_list) + start
|
|
if end < 0:
|
|
end = len(tags_list) + end
|
|
# clamp start and end values within list boundaries
|
|
start, end = min(start, len(tags_list)), min(end, len(tags_list))
|
|
start, end = max(start, 0), max(end, 0)
|
|
# merge all
|
|
for i in range(start, end):
|
|
output[i] = tags_list[i]
|
|
return ", ".join(list(output.values()))
|
|
|
|
def append_lora_name_if_empty(tags_list, lora_path, enabled):
|
|
if not enabled or len(tags_list) > 0:
|
|
return tags_list
|
|
filename = os.path.splitext(lora_path)[0]
|
|
filename = os.path.basename(filename)
|
|
|
|
tags_list.append(filename)
|
|
return tags_list
|