Remove FastAbsoluteSaver node (moved to ComfyUI-JSON-Dynamic)
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
@@ -1,19 +1,16 @@
|
|||||||
from .sharp_node import SharpnessAnalyzer, SharpFrameSelector
|
from .sharp_node import SharpnessAnalyzer, SharpFrameSelector
|
||||||
from .parallel_loader import ParallelSharpnessLoader
|
from .parallel_loader import ParallelSharpnessLoader
|
||||||
from .fast_saver import FastAbsoluteSaver # <--- Added this missing import
|
|
||||||
|
|
||||||
NODE_CLASS_MAPPINGS = {
|
NODE_CLASS_MAPPINGS = {
|
||||||
"SharpnessAnalyzer": SharpnessAnalyzer,
|
"SharpnessAnalyzer": SharpnessAnalyzer,
|
||||||
"SharpFrameSelector": SharpFrameSelector,
|
"SharpFrameSelector": SharpFrameSelector,
|
||||||
"ParallelSharpnessLoader": ParallelSharpnessLoader,
|
"ParallelSharpnessLoader": ParallelSharpnessLoader,
|
||||||
"FastAbsoluteSaver": FastAbsoluteSaver
|
|
||||||
}
|
}
|
||||||
|
|
||||||
NODE_DISPLAY_NAME_MAPPINGS = {
|
NODE_DISPLAY_NAME_MAPPINGS = {
|
||||||
"SharpnessAnalyzer": "1. Sharpness Analyzer",
|
"SharpnessAnalyzer": "1. Sharpness Analyzer",
|
||||||
"SharpFrameSelector": "2. Sharp Frame Selector",
|
"SharpFrameSelector": "2. Sharp Frame Selector",
|
||||||
"ParallelSharpnessLoader": "3. Parallel Video Loader (Sharpness)",
|
"ParallelSharpnessLoader": "3. Parallel Video Loader (Sharpness)",
|
||||||
"FastAbsoluteSaver": "Fast Absolute Saver (Metadata)"
|
|
||||||
}
|
}
|
||||||
|
|
||||||
__all__ = ["NODE_CLASS_MAPPINGS", "NODE_DISPLAY_NAME_MAPPINGS"]
|
__all__ = ["NODE_CLASS_MAPPINGS", "NODE_DISPLAY_NAME_MAPPINGS"]
|
||||||
431
fast_saver.py
431
fast_saver.py
@@ -1,431 +0,0 @@
|
|||||||
import os
|
|
||||||
import sys
|
|
||||||
import platform
|
|
||||||
import torch
|
|
||||||
import numpy as np
|
|
||||||
from PIL import Image, ExifTags
|
|
||||||
from PIL.PngImagePlugin import PngInfo
|
|
||||||
import concurrent.futures
|
|
||||||
import re
|
|
||||||
import time
|
|
||||||
import glob
|
|
||||||
import json
|
|
||||||
import subprocess
|
|
||||||
import shutil
|
|
||||||
import stat
|
|
||||||
import tempfile
|
|
||||||
import urllib.request
|
|
||||||
import zipfile
|
|
||||||
import tarfile
|
|
||||||
|
|
||||||
_NODE_DIR = os.path.dirname(os.path.abspath(__file__))
|
|
||||||
_FFMPEG_DIR = os.path.join(_NODE_DIR, "ffmpeg_bin")
|
|
||||||
|
|
||||||
|
|
||||||
def _get_ffmpeg():
|
|
||||||
"""Find or download a ffmpeg binary. Search order:
|
|
||||||
1. Bundled binary in this node's ffmpeg_bin/ folder
|
|
||||||
2. imageio_ffmpeg (shipped by VideoHelperSuite)
|
|
||||||
3. System PATH
|
|
||||||
4. Auto-download a static build into ffmpeg_bin/
|
|
||||||
"""
|
|
||||||
system = platform.system()
|
|
||||||
exe_name = "ffmpeg.exe" if system == "Windows" else "ffmpeg"
|
|
||||||
local_bin = os.path.join(_FFMPEG_DIR, exe_name)
|
|
||||||
|
|
||||||
# 1. Already downloaded
|
|
||||||
if os.path.isfile(local_bin):
|
|
||||||
return local_bin
|
|
||||||
|
|
||||||
# 2. imageio_ffmpeg
|
|
||||||
try:
|
|
||||||
import imageio_ffmpeg
|
|
||||||
path = imageio_ffmpeg.get_ffmpeg_exe()
|
|
||||||
if path and os.path.isfile(path):
|
|
||||||
return path
|
|
||||||
except Exception:
|
|
||||||
pass
|
|
||||||
|
|
||||||
# 3. System PATH
|
|
||||||
system_bin = shutil.which("ffmpeg")
|
|
||||||
if system_bin:
|
|
||||||
return system_bin
|
|
||||||
|
|
||||||
# 4. Auto-download static build
|
|
||||||
print("xx- FastSaver: ffmpeg not found. Downloading static build...")
|
|
||||||
os.makedirs(_FFMPEG_DIR, exist_ok=True)
|
|
||||||
|
|
||||||
urls = {
|
|
||||||
("Linux", "x86_64"): "https://johnvansickle.com/ffmpeg/releases/ffmpeg-release-amd64-static.tar.xz",
|
|
||||||
("Linux", "aarch64"): "https://johnvansickle.com/ffmpeg/releases/ffmpeg-release-arm64-static.tar.xz",
|
|
||||||
("Windows", "AMD64"): "https://www.gyan.dev/ffmpeg/builds/ffmpeg-release-essentials.zip",
|
|
||||||
}
|
|
||||||
|
|
||||||
machine = platform.machine()
|
|
||||||
key = (system, machine)
|
|
||||||
url = urls.get(key)
|
|
||||||
if not url:
|
|
||||||
raise RuntimeError(
|
|
||||||
f"No automatic ffmpeg download available for {system}/{machine}. "
|
|
||||||
f"Please install ffmpeg manually or place a binary in {_FFMPEG_DIR}"
|
|
||||||
)
|
|
||||||
|
|
||||||
archive_path = os.path.join(_FFMPEG_DIR, "ffmpeg_archive")
|
|
||||||
try:
|
|
||||||
urllib.request.urlretrieve(url, archive_path)
|
|
||||||
|
|
||||||
if url.endswith(".tar.xz"):
|
|
||||||
with tarfile.open(archive_path, "r:xz") as tar:
|
|
||||||
for member in tar.getmembers():
|
|
||||||
if member.name.endswith("/ffmpeg") and member.isfile():
|
|
||||||
member.name = exe_name
|
|
||||||
tar.extract(member, _FFMPEG_DIR)
|
|
||||||
break
|
|
||||||
elif url.endswith(".zip"):
|
|
||||||
with zipfile.ZipFile(archive_path, "r") as zf:
|
|
||||||
for name in zf.namelist():
|
|
||||||
if name.endswith("bin/ffmpeg.exe"):
|
|
||||||
data = zf.read(name)
|
|
||||||
with open(local_bin, "wb") as f:
|
|
||||||
f.write(data)
|
|
||||||
break
|
|
||||||
finally:
|
|
||||||
if os.path.exists(archive_path):
|
|
||||||
os.remove(archive_path)
|
|
||||||
|
|
||||||
if not os.path.isfile(local_bin):
|
|
||||||
raise RuntimeError(f"Failed to extract ffmpeg to {local_bin}")
|
|
||||||
|
|
||||||
# Make executable on Linux/Mac
|
|
||||||
if system != "Windows":
|
|
||||||
os.chmod(local_bin, os.stat(local_bin).st_mode | stat.S_IEXEC)
|
|
||||||
|
|
||||||
print(f"xx- FastSaver: ffmpeg downloaded to {local_bin}")
|
|
||||||
return local_bin
|
|
||||||
|
|
||||||
class FastAbsoluteSaver:
|
|
||||||
@classmethod
|
|
||||||
def INPUT_TYPES(s):
|
|
||||||
return {
|
|
||||||
"required": {
|
|
||||||
"images": ("IMAGE", ),
|
|
||||||
"output_path": ("STRING", {"default": "D:\\Datasets\\Sharp_Output"}),
|
|
||||||
"filename_prefix": ("STRING", {"default": "frame"}),
|
|
||||||
|
|
||||||
# --- FORMAT SWITCH ---
|
|
||||||
"save_format": (["png", "webp", "mp4", "webm"], ),
|
|
||||||
|
|
||||||
# --- NAMING CONTROL ---
|
|
||||||
"use_timestamp": ("BOOLEAN", {"default": False, "label": "Add Timestamp (Unique)"}),
|
|
||||||
"auto_increment": ("BOOLEAN", {"default": True, "label": "Auto-Increment Counter (Scan Folder)"}),
|
|
||||||
"counter_digits": ("INT", {"default": 4, "min": 1, "max": 12, "step": 1, "label": "Number Padding (000X)"}),
|
|
||||||
"filename_with_score": ("BOOLEAN", {"default": False, "label": "Append Score to Filename"}),
|
|
||||||
|
|
||||||
# --- METADATA & WORKFLOW ---
|
|
||||||
"metadata_key": ("STRING", {"default": "sharpness_score"}),
|
|
||||||
"save_workflow_metadata": ("BOOLEAN", {"default": False, "label": "Save ComfyUI Workflow (Graph)"}),
|
|
||||||
|
|
||||||
# --- PERFORMANCE ---
|
|
||||||
"max_threads": ("INT", {"default": 0, "min": 0, "max": 128, "step": 1, "label": "Max Threads (0=Auto)"}),
|
|
||||||
|
|
||||||
# --- WEBP SPECIFIC ---
|
|
||||||
"webp_lossless": ("BOOLEAN", {"default": True, "label": "WebP Lossless"}),
|
|
||||||
"webp_quality": ("INT", {"default": 100, "min": 0, "max": 100, "step": 1}),
|
|
||||||
"webp_method": ("INT", {"default": 4, "min": 0, "max": 6, "step": 1}),
|
|
||||||
|
|
||||||
# --- VIDEO SPECIFIC ---
|
|
||||||
"video_fps": ("INT", {"default": 24, "min": 1, "max": 120, "step": 1, "label": "Video FPS"}),
|
|
||||||
"video_crf": ("INT", {"default": 18, "min": 0, "max": 51, "step": 1, "label": "Video CRF (0=Lossless, 51=Worst)"}),
|
|
||||||
"video_pixel_format": (["yuv420p", "yuv444p"], {"label": "Pixel Format"}),
|
|
||||||
},
|
|
||||||
"optional": {
|
|
||||||
"scores_info": ("STRING", {"forceInput": True}),
|
|
||||||
},
|
|
||||||
# Hidden inputs used to capture the workflow graph
|
|
||||||
"hidden": {"prompt": "PROMPT", "extra_pnginfo": "EXTRA_PNGINFO"},
|
|
||||||
}
|
|
||||||
|
|
||||||
RETURN_TYPES = ()
|
|
||||||
FUNCTION = "save_images_fast"
|
|
||||||
OUTPUT_NODE = True
|
|
||||||
CATEGORY = "BetaHelper/IO"
|
|
||||||
|
|
||||||
def parse_info(self, info_str, batch_size):
|
|
||||||
if not info_str:
|
|
||||||
# No scores connected - return None for scores to indicate "not provided"
|
|
||||||
return ([0]*batch_size, None)
|
|
||||||
matches = re.findall(r"F:(\d+).*?Score:\s*(\d+(\.\d+)?)", info_str)
|
|
||||||
frames = []
|
|
||||||
scores = []
|
|
||||||
for m in matches:
|
|
||||||
try:
|
|
||||||
frames.append(int(m[0]))
|
|
||||||
scores.append(float(m[1]))
|
|
||||||
except ValueError:
|
|
||||||
pass
|
|
||||||
if len(frames) < batch_size:
|
|
||||||
missing = batch_size - len(frames)
|
|
||||||
frames.extend([0] * missing)
|
|
||||||
scores.extend([0.0] * missing)
|
|
||||||
return frames[:batch_size], scores[:batch_size]
|
|
||||||
|
|
||||||
def get_start_index(self, output_path, prefix):
|
|
||||||
# Scans the directory ONCE to find the highest existing number.
|
|
||||||
print(f"xx- FastSaver: Scanning folder for existing '{prefix}' files...")
|
|
||||||
files = glob.glob(os.path.join(output_path, f"{prefix}*.*"))
|
|
||||||
|
|
||||||
max_idx = 0
|
|
||||||
# Check specifically for prefix_NUMBER pattern to avoid confusing timestamps
|
|
||||||
pattern = re.compile(rf"{re.escape(prefix)}_(\d+)")
|
|
||||||
|
|
||||||
for f in files:
|
|
||||||
fname = os.path.basename(f)
|
|
||||||
match = pattern.match(fname)
|
|
||||||
if match:
|
|
||||||
try:
|
|
||||||
val = int(match.group(1))
|
|
||||||
if val > max_idx:
|
|
||||||
max_idx = val
|
|
||||||
except ValueError:
|
|
||||||
continue
|
|
||||||
print(f"xx- FastSaver: Found highest index {max_idx}. Starting at {max_idx + 1}")
|
|
||||||
return max_idx + 1
|
|
||||||
|
|
||||||
def save_single_image(self, tensor_img, full_path, score, key_name, fmt, lossless, quality, method,
|
|
||||||
save_workflow, prompt_data, extra_data):
|
|
||||||
try:
|
|
||||||
array = 255. * tensor_img.cpu().numpy()
|
|
||||||
img = Image.fromarray(np.clip(array, 0, 255).astype(np.uint8))
|
|
||||||
|
|
||||||
# --- METADATA PREPARATION ---
|
|
||||||
meta_png = PngInfo()
|
|
||||||
exif_bytes = None
|
|
||||||
|
|
||||||
# 1. Custom Score Metadata (only if score was actually provided)
|
|
||||||
if fmt == "png":
|
|
||||||
if score is not None:
|
|
||||||
meta_png.add_text(key_name, str(score))
|
|
||||||
meta_png.add_text("software", "ComfyUI_Parallel_Node")
|
|
||||||
|
|
||||||
# 2. ComfyUI Workflow Metadata (If requested)
|
|
||||||
if save_workflow:
|
|
||||||
# Prepare JSON payloads
|
|
||||||
workflow_json = json.dumps(extra_data.get("workflow", {})) if extra_data else "{}"
|
|
||||||
prompt_json = json.dumps(prompt_data) if prompt_data else "{}"
|
|
||||||
|
|
||||||
if fmt == "png":
|
|
||||||
# Standard PNG text chunks
|
|
||||||
meta_png.add_text("prompt", prompt_json)
|
|
||||||
meta_png.add_text("workflow", workflow_json)
|
|
||||||
|
|
||||||
elif fmt == "webp":
|
|
||||||
# WebP: Embed in Exif UserComment (Standard ComfyUI method)
|
|
||||||
# We construct a JSON dict containing the workflow
|
|
||||||
exif_payload = {
|
|
||||||
"prompt": prompt_data,
|
|
||||||
"workflow": extra_data.get("workflow", {}) if extra_data else {}
|
|
||||||
}
|
|
||||||
# We also add the custom score here for WebP readers that check Exif
|
|
||||||
if score is not None:
|
|
||||||
exif_payload[key_name] = score
|
|
||||||
|
|
||||||
user_comment = json.dumps(exif_payload)
|
|
||||||
|
|
||||||
# Create Exif data with tag 0x9286 (UserComment)
|
|
||||||
exif_dict = {
|
|
||||||
ExifTags.IFD.Exif: {
|
|
||||||
0x9286: user_comment.encode('utf-8')
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
# Pillow requires raw bytes for 'exif='
|
|
||||||
# Since we want to avoid 'piexif' dependency, we do a lightweight workaround:
|
|
||||||
# We just save the image. Pillow WebP writer doesn't support easy Exif writing
|
|
||||||
# without external libs or pre-existing exif.
|
|
||||||
#
|
|
||||||
# FALLBACK: If we can't write complex Exif easily without piexif,
|
|
||||||
# we will skip WebP workflow embedding to keep this node dependency-free.
|
|
||||||
#
|
|
||||||
# BUT, ComfyUI users expect it.
|
|
||||||
# Strategy: If format is WebP and workflow is ON, we assume
|
|
||||||
# the user is okay with a slightly slower save or we skip it if dependencies missing.
|
|
||||||
# For this "Fast" node, we will skip the complex Exif write to prevent errors/bloat.
|
|
||||||
pass
|
|
||||||
|
|
||||||
# --- SAVING ---
|
|
||||||
if fmt == "png":
|
|
||||||
img.save(full_path, format="PNG", pnginfo=meta_png, compress_level=1)
|
|
||||||
|
|
||||||
elif fmt == "webp":
|
|
||||||
img.save(full_path, format="WEBP", lossless=lossless, quality=quality, method=method)
|
|
||||||
|
|
||||||
return True
|
|
||||||
except Exception as e:
|
|
||||||
print(f"xx- Error saving {full_path}: {e}")
|
|
||||||
return False
|
|
||||||
|
|
||||||
def save_video(self, images, output_path, filename_prefix, use_timestamp, fps, crf, pixel_format, video_format,
|
|
||||||
scores_list=None, metadata_key="sharpness_score", save_workflow=False, prompt_data=None, extra_data=None):
|
|
||||||
"""Save image batch as a video file using ffmpeg."""
|
|
||||||
ffmpeg_path = _get_ffmpeg()
|
|
||||||
|
|
||||||
ts_str = f"_{int(time.time())}" if use_timestamp else ""
|
|
||||||
ext = ".mp4" if video_format == "mp4" else ".webm"
|
|
||||||
out_file = os.path.join(output_path, f"{filename_prefix}{ts_str}{ext}")
|
|
||||||
|
|
||||||
batch_size = len(images)
|
|
||||||
h, w = images[0].shape[0], images[0].shape[1]
|
|
||||||
|
|
||||||
# --- BUILD METADATA FILE (avoids arg-too-long for large workflows) ---
|
|
||||||
meta_lines = [";FFMETADATA1"]
|
|
||||||
meta_lines.append("software=ComfyUI_FastAbsoluteSaver")
|
|
||||||
|
|
||||||
if scores_list:
|
|
||||||
avg_score = sum(scores_list) / len(scores_list)
|
|
||||||
meta_lines.append(f"{metadata_key}_avg={avg_score:.2f}")
|
|
||||||
meta_lines.append(f"{metadata_key}_all={','.join(f'{s:.2f}' for s in scores_list)}")
|
|
||||||
|
|
||||||
if save_workflow:
|
|
||||||
if prompt_data:
|
|
||||||
# Escape ffmetadata special chars: =, ;, #, \ and newlines
|
|
||||||
prompt_str = json.dumps(prompt_data).replace("\\", "\\\\").replace("=", "\\=").replace(";", "\\;").replace("#", "\\#").replace("\n", "\\\n")
|
|
||||||
meta_lines.append(f"prompt={prompt_str}")
|
|
||||||
if extra_data:
|
|
||||||
workflow = extra_data.get("workflow", {})
|
|
||||||
if workflow:
|
|
||||||
workflow_str = json.dumps(workflow).replace("\\", "\\\\").replace("=", "\\=").replace(";", "\\;").replace("#", "\\#").replace("\n", "\\\n")
|
|
||||||
meta_lines.append(f"workflow={workflow_str}")
|
|
||||||
|
|
||||||
self._meta_tmpfile = tempfile.NamedTemporaryFile(mode="w", suffix=".txt", delete=False, encoding="utf-8")
|
|
||||||
self._meta_tmpfile.write("\n".join(meta_lines))
|
|
||||||
self._meta_tmpfile.close()
|
|
||||||
meta_file = self._meta_tmpfile.name
|
|
||||||
|
|
||||||
if video_format == "mp4":
|
|
||||||
codec = "libx264"
|
|
||||||
cmd = [
|
|
||||||
ffmpeg_path, "-y",
|
|
||||||
"-f", "rawvideo", "-pix_fmt", "rgb24",
|
|
||||||
"-s", f"{w}x{h}", "-r", str(fps),
|
|
||||||
"-i", "-",
|
|
||||||
"-i", meta_file, "-map_metadata", "1",
|
|
||||||
"-c:v", codec, "-crf", str(crf),
|
|
||||||
"-pix_fmt", pixel_format,
|
|
||||||
"-movflags", "+faststart",
|
|
||||||
out_file
|
|
||||||
]
|
|
||||||
else: # webm
|
|
||||||
codec = "libvpx-vp9"
|
|
||||||
cmd = [
|
|
||||||
ffmpeg_path, "-y",
|
|
||||||
"-f", "rawvideo", "-pix_fmt", "rgb24",
|
|
||||||
"-s", f"{w}x{h}", "-r", str(fps),
|
|
||||||
"-i", "-",
|
|
||||||
"-i", meta_file, "-map_metadata", "1",
|
|
||||||
"-c:v", codec, "-crf", str(crf), "-b:v", "0",
|
|
||||||
"-pix_fmt", pixel_format,
|
|
||||||
out_file
|
|
||||||
]
|
|
||||||
|
|
||||||
print(f"xx- FastSaver: Encoding {batch_size} frames to {out_file} ({codec}, crf={crf}, {fps}fps)...")
|
|
||||||
|
|
||||||
proc = subprocess.Popen(cmd, stdin=subprocess.PIPE, stderr=subprocess.PIPE)
|
|
||||||
try:
|
|
||||||
for img_tensor in images:
|
|
||||||
frame = (255.0 * img_tensor.cpu().numpy()).clip(0, 255).astype(np.uint8)
|
|
||||||
proc.stdin.write(frame.tobytes())
|
|
||||||
proc.stdin.close()
|
|
||||||
except BrokenPipeError:
|
|
||||||
pass
|
|
||||||
stderr = proc.stderr.read()
|
|
||||||
proc.wait()
|
|
||||||
|
|
||||||
# Clean up metadata temp file
|
|
||||||
try:
|
|
||||||
os.remove(meta_file)
|
|
||||||
except OSError:
|
|
||||||
pass
|
|
||||||
|
|
||||||
if proc.returncode != 0:
|
|
||||||
raise RuntimeError(f"ffmpeg failed: {stderr.decode()}")
|
|
||||||
|
|
||||||
print(f"xx- FastSaver: Video saved to {out_file}")
|
|
||||||
return out_file
|
|
||||||
|
|
||||||
def save_images_fast(self, images, output_path, filename_prefix, save_format, use_timestamp, auto_increment, counter_digits,
|
|
||||||
max_threads, filename_with_score, metadata_key, save_workflow_metadata,
|
|
||||||
webp_lossless, webp_quality, webp_method,
|
|
||||||
video_fps, video_crf, video_pixel_format,
|
|
||||||
scores_info=None, prompt=None, extra_pnginfo=None):
|
|
||||||
|
|
||||||
output_path = output_path.strip('"')
|
|
||||||
if not os.path.exists(output_path):
|
|
||||||
try:
|
|
||||||
os.makedirs(output_path, exist_ok=True)
|
|
||||||
except OSError:
|
|
||||||
raise ValueError(f"Could not create directory: {output_path}")
|
|
||||||
|
|
||||||
if images is None or len(images) == 0:
|
|
||||||
raise ValueError("No images provided to FastAbsoluteSaver.")
|
|
||||||
|
|
||||||
# --- VIDEO PATH (check early, before image-specific logic) ---
|
|
||||||
if save_format in ("mp4", "webm"):
|
|
||||||
batch_size = len(images)
|
|
||||||
_, scores_list = self.parse_info(scores_info, batch_size)
|
|
||||||
self.save_video(images, output_path, filename_prefix, use_timestamp,
|
|
||||||
video_fps, video_crf, video_pixel_format, save_format,
|
|
||||||
scores_list=scores_list, metadata_key=metadata_key,
|
|
||||||
save_workflow=save_workflow_metadata, prompt_data=prompt,
|
|
||||||
extra_data=extra_pnginfo)
|
|
||||||
return {"ui": {"images": []}}
|
|
||||||
|
|
||||||
if max_threads == 0:
|
|
||||||
max_threads = os.cpu_count() or 4
|
|
||||||
|
|
||||||
batch_size = len(images)
|
|
||||||
frame_indices, scores_list = self.parse_info(scores_info, batch_size)
|
|
||||||
|
|
||||||
# --- INDEX LOGIC ---
|
|
||||||
start_counter = 0
|
|
||||||
using_real_frames = any(idx > 0 for idx in frame_indices)
|
|
||||||
|
|
||||||
if auto_increment and not use_timestamp and not using_real_frames:
|
|
||||||
start_counter = self.get_start_index(output_path, filename_prefix)
|
|
||||||
|
|
||||||
ts_str = f"_{int(time.time())}" if use_timestamp else ""
|
|
||||||
|
|
||||||
print(f"xx- FastSaver: Saving {batch_size} images to {output_path}...")
|
|
||||||
|
|
||||||
with concurrent.futures.ThreadPoolExecutor(max_workers=max_threads) as executor:
|
|
||||||
futures = []
|
|
||||||
|
|
||||||
for i, img_tensor in enumerate(images):
|
|
||||||
real_frame_num = frame_indices[i]
|
|
||||||
current_score = scores_list[i] if scores_list else None
|
|
||||||
|
|
||||||
if real_frame_num > 0:
|
|
||||||
number_part = real_frame_num
|
|
||||||
else:
|
|
||||||
number_part = start_counter + i
|
|
||||||
|
|
||||||
fmt_str = f"{{:0{counter_digits}d}}"
|
|
||||||
number_str = fmt_str.format(number_part)
|
|
||||||
|
|
||||||
base_name = f"{filename_prefix}{ts_str}_{number_str}"
|
|
||||||
|
|
||||||
if filename_with_score and current_score is not None:
|
|
||||||
base_name += f"_{int(current_score)}"
|
|
||||||
|
|
||||||
ext = ".webp" if save_format == "webp" else ".png"
|
|
||||||
full_path = os.path.join(output_path, f"{base_name}{ext}")
|
|
||||||
|
|
||||||
futures.append(executor.submit(
|
|
||||||
self.save_single_image,
|
|
||||||
img_tensor, full_path, current_score, metadata_key,
|
|
||||||
save_format, webp_lossless, webp_quality, webp_method,
|
|
||||||
save_workflow_metadata, prompt, extra_pnginfo
|
|
||||||
))
|
|
||||||
|
|
||||||
concurrent.futures.wait(futures)
|
|
||||||
|
|
||||||
return {"ui": {"images": []}}
|
|
||||||
Reference in New Issue
Block a user