Merge pull request 'global-button' (#1) from global-button into main
Reviewed-on: #1
This commit was merged in pull request #1.
This commit is contained in:
128
engine.py
128
engine.py
@@ -10,30 +10,24 @@ class SorterEngine:
|
|||||||
# --- 1. DATABASE INITIALIZATION ---
|
# --- 1. DATABASE INITIALIZATION ---
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def init_db():
|
def init_db():
|
||||||
"""Initializes all SQLite tables for the multi-tab system."""
|
"""Initializes tables, including the new HISTORY log."""
|
||||||
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
|
|
||||||
# Profiles Table: 9 columns for independent tab paths
|
# Existing tables...
|
||||||
cursor.execute('''CREATE TABLE IF NOT EXISTS profiles
|
cursor.execute('''CREATE TABLE IF NOT EXISTS profiles
|
||||||
(name TEXT PRIMARY KEY,
|
(name TEXT PRIMARY KEY, tab1_target TEXT, tab2_target TEXT, tab2_control TEXT,
|
||||||
tab1_target TEXT,
|
tab4_source TEXT, tab4_out TEXT, mode TEXT, tab5_source TEXT, tab5_out TEXT)''')
|
||||||
tab2_target TEXT, tab2_control TEXT,
|
|
||||||
tab4_source TEXT, tab4_out TEXT,
|
|
||||||
mode TEXT,
|
|
||||||
tab5_source TEXT, tab5_out TEXT)''')
|
|
||||||
|
|
||||||
cursor.execute('''CREATE TABLE IF NOT EXISTS folder_ids (path TEXT PRIMARY KEY, folder_id INTEGER)''')
|
cursor.execute('''CREATE TABLE IF NOT EXISTS folder_ids (path TEXT PRIMARY KEY, folder_id INTEGER)''')
|
||||||
cursor.execute('''CREATE TABLE IF NOT EXISTS categories (name TEXT PRIMARY KEY)''')
|
cursor.execute('''CREATE TABLE IF NOT EXISTS categories (name TEXT PRIMARY KEY)''')
|
||||||
|
|
||||||
# Staging Area: Tracks pending renames for the Gallery Tab
|
|
||||||
cursor.execute('''CREATE TABLE IF NOT EXISTS staging_area
|
cursor.execute('''CREATE TABLE IF NOT EXISTS staging_area
|
||||||
(original_path TEXT PRIMARY KEY,
|
(original_path TEXT PRIMARY KEY, target_category TEXT, new_name TEXT, is_marked INTEGER DEFAULT 0)''')
|
||||||
target_category TEXT,
|
|
||||||
new_name TEXT,
|
# --- NEW: HISTORY TABLE ---
|
||||||
is_marked INTEGER DEFAULT 0)''')
|
cursor.execute('''CREATE TABLE IF NOT EXISTS processed_log
|
||||||
|
(source_path TEXT PRIMARY KEY, category TEXT, action_type TEXT)''')
|
||||||
|
|
||||||
# Seed default categories
|
# Seed categories if empty
|
||||||
cursor.execute("SELECT COUNT(*) FROM categories")
|
cursor.execute("SELECT COUNT(*) FROM categories")
|
||||||
if cursor.fetchone()[0] == 0:
|
if cursor.fetchone()[0] == 0:
|
||||||
for cat in ["_TRASH", "Default", "Action", "Solo"]:
|
for cat in ["_TRASH", "Default", "Action", "Solo"]:
|
||||||
@@ -69,6 +63,31 @@ class SorterEngine:
|
|||||||
cursor.execute("INSERT OR REPLACE INTO profiles VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?)", new_values)
|
cursor.execute("INSERT OR REPLACE INTO profiles VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?)", new_values)
|
||||||
conn.commit()
|
conn.commit()
|
||||||
conn.close()
|
conn.close()
|
||||||
|
@staticmethod
|
||||||
|
def load_batch_parallel(image_paths, quality):
|
||||||
|
"""
|
||||||
|
Multithreaded loader: Compresses multiple images in parallel.
|
||||||
|
Returns a dictionary {path: bytes_io}
|
||||||
|
"""
|
||||||
|
import concurrent.futures
|
||||||
|
|
||||||
|
results = {}
|
||||||
|
|
||||||
|
# Helper function to run in thread
|
||||||
|
def process_one(path):
|
||||||
|
return path, SorterEngine.compress_for_web(path, quality)
|
||||||
|
|
||||||
|
# Use ThreadPool to parallelize IO-bound tasks
|
||||||
|
with concurrent.futures.ThreadPoolExecutor(max_workers=8) as executor:
|
||||||
|
# Submit all tasks
|
||||||
|
future_to_path = {executor.submit(process_one, p): p for p in image_paths}
|
||||||
|
|
||||||
|
# Gather results as they complete
|
||||||
|
for future in concurrent.futures.as_completed(future_to_path):
|
||||||
|
path, data = future.result()
|
||||||
|
results[path] = data
|
||||||
|
|
||||||
|
return results
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def load_profiles():
|
def load_profiles():
|
||||||
@@ -234,22 +253,19 @@ class SorterEngine:
|
|||||||
return {r[0]: {"cat": r[1], "name": r[2], "marked": r[3]} for r in rows}
|
return {r[0]: {"cat": r[1], "name": r[2], "marked": r[3]} for r in rows}
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def commit_staging(output_root, cleanup_mode, source_root=None):
|
def commit_global(output_root, cleanup_mode, operation="Move", source_root=None):
|
||||||
"""Global commit directly to output root (No Subfolders)."""
|
"""Commits ALL staged files (Global Apply)."""
|
||||||
data = SorterEngine.get_staged_data()
|
data = SorterEngine.get_staged_data()
|
||||||
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
staged_paths = set(data.keys())
|
|
||||||
|
|
||||||
if not os.path.exists(output_root):
|
if not os.path.exists(output_root): os.makedirs(output_root, exist_ok=True)
|
||||||
os.makedirs(output_root, exist_ok=True)
|
|
||||||
|
|
||||||
|
# 1. Process all Staged Items
|
||||||
for old_p, info in data.items():
|
for old_p, info in data.items():
|
||||||
if os.path.exists(old_p):
|
if os.path.exists(old_p):
|
||||||
# CHANGED: Direct move to root
|
|
||||||
final_dst = os.path.join(output_root, info['name'])
|
final_dst = os.path.join(output_root, info['name'])
|
||||||
|
|
||||||
# Collision Safety for global commit
|
|
||||||
if os.path.exists(final_dst):
|
if os.path.exists(final_dst):
|
||||||
root, ext = os.path.splitext(info['name'])
|
root, ext = os.path.splitext(info['name'])
|
||||||
c = 1
|
c = 1
|
||||||
@@ -257,17 +273,28 @@ class SorterEngine:
|
|||||||
final_dst = os.path.join(output_root, f"{root}_{c}{ext}")
|
final_dst = os.path.join(output_root, f"{root}_{c}{ext}")
|
||||||
c += 1
|
c += 1
|
||||||
|
|
||||||
shutil.move(old_p, final_dst)
|
if operation == "Copy":
|
||||||
|
shutil.copy2(old_p, final_dst)
|
||||||
|
else:
|
||||||
|
shutil.move(old_p, final_dst)
|
||||||
|
|
||||||
|
# Log History
|
||||||
|
cursor.execute("INSERT OR REPLACE INTO processed_log VALUES (?, ?, ?)",
|
||||||
|
(old_p, info['cat'], operation))
|
||||||
|
|
||||||
|
# 2. Global Cleanup (Optional)
|
||||||
|
# Only run cleanup if explicitly asked, as global cleanup is risky
|
||||||
if cleanup_mode != "Keep" and source_root:
|
if cleanup_mode != "Keep" and source_root:
|
||||||
for img_p in SorterEngine.get_images(source_root, recursive=True):
|
all_imgs = SorterEngine.get_images(source_root, recursive=True)
|
||||||
if img_p not in staged_paths:
|
for img_p in all_imgs:
|
||||||
|
if img_p not in data: # Not currently staged
|
||||||
if cleanup_mode == "Move to Unused":
|
if cleanup_mode == "Move to Unused":
|
||||||
un_dir = os.path.join(source_root, "unused")
|
unused_dir = os.path.join(source_root, "unused")
|
||||||
os.makedirs(un_dir, exist_ok=True)
|
os.makedirs(unused_dir, exist_ok=True)
|
||||||
shutil.move(img_p, os.path.join(un_dir, os.path.basename(img_p)))
|
shutil.move(img_p, os.path.join(unused_dir, os.path.basename(img_p)))
|
||||||
elif cleanup_mode == "Delete": os.remove(img_p)
|
elif cleanup_mode == "Delete":
|
||||||
|
os.remove(img_p)
|
||||||
|
|
||||||
cursor.execute("DELETE FROM staging_area")
|
cursor.execute("DELETE FROM staging_area")
|
||||||
conn.commit()
|
conn.commit()
|
||||||
conn.close()
|
conn.close()
|
||||||
@@ -346,28 +373,34 @@ class SorterEngine:
|
|||||||
if 'c_dst' in action and os.path.exists(action['c_dst']):
|
if 'c_dst' in action and os.path.exists(action['c_dst']):
|
||||||
shutil.move(action['c_dst'], action['c_src'])
|
shutil.move(action['c_dst'], action['c_src'])
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def get_processed_log():
|
||||||
|
"""Retrieves history of processed files."""
|
||||||
|
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
||||||
|
cursor = conn.cursor()
|
||||||
|
cursor.execute("SELECT * FROM processed_log")
|
||||||
|
rows = cursor.fetchall()
|
||||||
|
conn.close()
|
||||||
|
return {r[0]: {"cat": r[1], "action": r[2]} for r in rows}
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def commit_batch(file_list, output_root, cleanup_mode, operation="Move"):
|
def commit_batch(file_list, output_root, cleanup_mode, operation="Move"):
|
||||||
"""
|
"""Commits specified files and LOGS them to history."""
|
||||||
Commits files with support for both MOVE and COPY.
|
|
||||||
"""
|
|
||||||
data = SorterEngine.get_staged_data()
|
data = SorterEngine.get_staged_data()
|
||||||
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
||||||
cursor = conn.cursor()
|
cursor = conn.cursor()
|
||||||
|
|
||||||
# Ensure output root exists
|
if not os.path.exists(output_root): os.makedirs(output_root, exist_ok=True)
|
||||||
if not os.path.exists(output_root):
|
|
||||||
os.makedirs(output_root, exist_ok=True)
|
|
||||||
|
|
||||||
for file_path in file_list:
|
for file_path in file_list:
|
||||||
if not os.path.exists(file_path): continue
|
if not os.path.exists(file_path): continue
|
||||||
|
|
||||||
# --- CASE A: File is TAGGED ---
|
# --- CASE A: Tagged ---
|
||||||
if file_path in data and data[file_path]['marked']:
|
if file_path in data and data[file_path]['marked']:
|
||||||
info = data[file_path]
|
info = data[file_path]
|
||||||
final_dst = os.path.join(output_root, info['name'])
|
final_dst = os.path.join(output_root, info['name'])
|
||||||
|
|
||||||
# Collision Safety
|
# Collision Check
|
||||||
if os.path.exists(final_dst):
|
if os.path.exists(final_dst):
|
||||||
root, ext = os.path.splitext(info['name'])
|
root, ext = os.path.splitext(info['name'])
|
||||||
c = 1
|
c = 1
|
||||||
@@ -375,22 +408,21 @@ class SorterEngine:
|
|||||||
final_dst = os.path.join(output_root, f"{root}_{c}{ext}")
|
final_dst = os.path.join(output_root, f"{root}_{c}{ext}")
|
||||||
c += 1
|
c += 1
|
||||||
|
|
||||||
# OPERATION CHECK: Move vs Copy
|
# Action
|
||||||
if operation == "Copy":
|
if operation == "Copy":
|
||||||
shutil.copy2(file_path, final_dst) # copy2 preserves metadata
|
shutil.copy2(file_path, final_dst)
|
||||||
else:
|
else:
|
||||||
shutil.move(file_path, final_dst)
|
shutil.move(file_path, final_dst)
|
||||||
|
|
||||||
# Remove from staging database
|
# Update DB: Remove from Staging, Add to History
|
||||||
cursor.execute("DELETE FROM staging_area WHERE original_path = ?", (file_path,))
|
cursor.execute("DELETE FROM staging_area WHERE original_path = ?", (file_path,))
|
||||||
|
cursor.execute("INSERT OR REPLACE INTO processed_log VALUES (?, ?, ?)",
|
||||||
|
(file_path, info['cat'], operation))
|
||||||
|
|
||||||
# --- CASE B: File is UNTAGGED (Cleanup) ---
|
# --- CASE B: Cleanup ---
|
||||||
# Note: If we COPIED a tagged file, the original stays in source,
|
|
||||||
# but this 'elif' ensures we don't accidentally delete it as 'untagged'.
|
|
||||||
elif cleanup_mode != "Keep":
|
elif cleanup_mode != "Keep":
|
||||||
if cleanup_mode == "Move to Unused":
|
if cleanup_mode == "Move to Unused":
|
||||||
parent = os.path.dirname(file_path)
|
unused_dir = os.path.join(os.path.dirname(file_path), "unused")
|
||||||
unused_dir = os.path.join(parent, "unused")
|
|
||||||
os.makedirs(unused_dir, exist_ok=True)
|
os.makedirs(unused_dir, exist_ok=True)
|
||||||
shutil.move(file_path, os.path.join(unused_dir, os.path.basename(file_path)))
|
shutil.move(file_path, os.path.join(unused_dir, os.path.basename(file_path)))
|
||||||
elif cleanup_mode == "Delete":
|
elif cleanup_mode == "Delete":
|
||||||
|
|||||||
@@ -4,57 +4,69 @@ import math
|
|||||||
from engine import SorterEngine
|
from engine import SorterEngine
|
||||||
|
|
||||||
# ==========================================
|
# ==========================================
|
||||||
# 1. GLOBAL CALLBACKS (Prevents Page Refresh)
|
# 1. CACHED DATA LOADER (The Fix)
|
||||||
# ==========================================
|
# ==========================================
|
||||||
|
@st.cache_data(show_spinner=False)
|
||||||
|
def get_cached_images(path, mutation_id):
|
||||||
|
"""
|
||||||
|
Scans the folder ONLY when 'path' or 'mutation_id' changes.
|
||||||
|
Navigating pages does NOT change these, so it remains instant.
|
||||||
|
"""
|
||||||
|
return SorterEngine.get_images(path, recursive=True)
|
||||||
|
|
||||||
|
|
||||||
|
# ==========================================
|
||||||
|
# 2. CALLBACKS (Updated with Refresh Logic)
|
||||||
|
# ==========================================
|
||||||
|
def trigger_refresh():
|
||||||
|
"""Increments the mutation counter to force a file re-scan."""
|
||||||
|
if 't5_file_id' not in st.session_state: st.session_state.t5_file_id = 0
|
||||||
|
st.session_state.t5_file_id += 1
|
||||||
|
|
||||||
def cb_tag_image(img_path, selected_cat):
|
def cb_tag_image(img_path, selected_cat):
|
||||||
"""Tags an image. Updates DB immediately."""
|
|
||||||
if selected_cat.startswith("---") or selected_cat == "":
|
if selected_cat.startswith("---") or selected_cat == "":
|
||||||
st.toast("⚠️ Select a valid category first!", icon="🚫")
|
st.toast("⚠️ Select a valid category first!", icon="🚫")
|
||||||
return
|
return
|
||||||
|
|
||||||
staged = SorterEngine.get_staged_data()
|
staged = SorterEngine.get_staged_data()
|
||||||
ext = os.path.splitext(img_path)[1]
|
ext = os.path.splitext(img_path)[1]
|
||||||
|
|
||||||
# Auto-increment logic
|
|
||||||
count = len([v for v in staged.values() if v['cat'] == selected_cat]) + 1
|
count = len([v for v in staged.values() if v['cat'] == selected_cat]) + 1
|
||||||
new_name = f"{selected_cat}_{count:03d}{ext}"
|
new_name = f"{selected_cat}_{count:03d}{ext}"
|
||||||
|
|
||||||
SorterEngine.stage_image(img_path, selected_cat, new_name)
|
SorterEngine.stage_image(img_path, selected_cat, new_name)
|
||||||
|
# Note: Tagging does NOT need a file re-scan, just a grid refresh.
|
||||||
|
|
||||||
def cb_untag_image(img_path):
|
def cb_untag_image(img_path):
|
||||||
"""Untags an image."""
|
|
||||||
SorterEngine.clear_staged_item(img_path)
|
SorterEngine.clear_staged_item(img_path)
|
||||||
|
|
||||||
def cb_delete_image(img_path):
|
def cb_delete_image(img_path):
|
||||||
"""Moves image to trash."""
|
|
||||||
SorterEngine.delete_to_trash(img_path)
|
SorterEngine.delete_to_trash(img_path)
|
||||||
|
trigger_refresh() # Force re-scan so the image disappears from the list
|
||||||
|
|
||||||
def cb_apply_batch(current_batch, path_o, cleanup_mode, operation):
|
def cb_apply_batch(current_batch, path_o, cleanup_mode, operation):
|
||||||
"""Commits the batch with the specified operation (Move/Copy)."""
|
|
||||||
SorterEngine.commit_batch(current_batch, path_o, cleanup_mode, operation)
|
SorterEngine.commit_batch(current_batch, path_o, cleanup_mode, operation)
|
||||||
|
trigger_refresh() # Force re-scan to remove moved files
|
||||||
|
|
||||||
|
def cb_apply_global(path_o, cleanup_mode, operation, path_s):
|
||||||
|
SorterEngine.commit_global(path_o, cleanup_mode, operation, source_root=path_s)
|
||||||
|
trigger_refresh() # Force re-scan
|
||||||
|
|
||||||
def cb_change_page(delta):
|
def cb_change_page(delta):
|
||||||
"""Updates page number (-1 or +1)."""
|
if 't5_page' not in st.session_state: st.session_state.t5_page = 0
|
||||||
if 't5_page' not in st.session_state:
|
|
||||||
st.session_state.t5_page = 0
|
|
||||||
st.session_state.t5_page += delta
|
st.session_state.t5_page += delta
|
||||||
|
# No trigger_refresh() here -> This is why page turning is now instant!
|
||||||
|
|
||||||
def cb_jump_page(k):
|
def cb_jump_page(k):
|
||||||
"""Updates page number from direct input box."""
|
|
||||||
val = st.session_state[k]
|
val = st.session_state[k]
|
||||||
st.session_state.t5_page = val - 1
|
st.session_state.t5_page = val - 1
|
||||||
|
|
||||||
|
|
||||||
# ==========================================
|
# ==========================================
|
||||||
# 2. FRAGMENT: SIDEBAR (Category Manager)
|
# 3. FRAGMENTS (Sidebar, Grid, Batch)
|
||||||
# ==========================================
|
# ==========================================
|
||||||
|
# ... (Sidebar code remains exactly the same) ...
|
||||||
@st.fragment
|
@st.fragment
|
||||||
def render_sidebar_content():
|
def render_sidebar_content():
|
||||||
st.divider()
|
st.divider()
|
||||||
st.subheader("🏷️ Category Manager")
|
st.subheader("🏷️ Category Manager")
|
||||||
|
|
||||||
# --- PREPARE LIST (With Separators) ---
|
|
||||||
cats = SorterEngine.get_categories()
|
cats = SorterEngine.get_categories()
|
||||||
processed_cats = []
|
processed_cats = []
|
||||||
last_char = ""
|
last_char = ""
|
||||||
@@ -66,26 +78,16 @@ def render_sidebar_content():
|
|||||||
processed_cats.append(cat)
|
processed_cats.append(cat)
|
||||||
last_char = current_char
|
last_char = current_char
|
||||||
|
|
||||||
# --- STATE SYNC ---
|
if "t5_active_cat" not in st.session_state: st.session_state.t5_active_cat = cats[0] if cats else "Default"
|
||||||
if "t5_active_cat" not in st.session_state:
|
|
||||||
st.session_state.t5_active_cat = cats[0] if cats else "Default"
|
|
||||||
|
|
||||||
# Fallback if selection was deleted
|
|
||||||
current_selection = st.session_state.t5_active_cat
|
current_selection = st.session_state.t5_active_cat
|
||||||
if not current_selection.startswith("---") and current_selection not in cats:
|
if not current_selection.startswith("---") and current_selection not in cats:
|
||||||
st.session_state.t5_active_cat = cats[0] if cats else "Default"
|
st.session_state.t5_active_cat = cats[0] if cats else "Default"
|
||||||
|
|
||||||
# --- RADIO SELECTION ---
|
|
||||||
selection = st.radio("Active Tag", processed_cats, key="t5_radio_select")
|
selection = st.radio("Active Tag", processed_cats, key="t5_radio_select")
|
||||||
|
if not selection.startswith("---"): st.session_state.t5_active_cat = selection
|
||||||
if not selection.startswith("---"):
|
|
||||||
st.session_state.t5_active_cat = selection
|
|
||||||
|
|
||||||
st.divider()
|
st.divider()
|
||||||
|
|
||||||
# --- TABS: ADD / EDIT ---
|
|
||||||
tab_add, tab_edit = st.tabs(["➕ Add", "✏️ Edit"])
|
tab_add, tab_edit = st.tabs(["➕ Add", "✏️ Edit"])
|
||||||
|
|
||||||
with tab_add:
|
with tab_add:
|
||||||
c1, c2 = st.columns([3, 1])
|
c1, c2 = st.columns([3, 1])
|
||||||
new_cat = c1.text_input("New Name", label_visibility="collapsed", placeholder="New...", key="t5_new_cat")
|
new_cat = c1.text_input("New Name", label_visibility="collapsed", placeholder="New...", key="t5_new_cat")
|
||||||
@@ -93,119 +95,101 @@ def render_sidebar_content():
|
|||||||
if new_cat:
|
if new_cat:
|
||||||
SorterEngine.add_category(new_cat)
|
SorterEngine.add_category(new_cat)
|
||||||
st.rerun()
|
st.rerun()
|
||||||
|
|
||||||
with tab_edit:
|
with tab_edit:
|
||||||
target_cat = st.session_state.t5_active_cat
|
target_cat = st.session_state.t5_active_cat
|
||||||
is_valid = target_cat and not target_cat.startswith("---") and target_cat in cats
|
if target_cat and not target_cat.startswith("---") and target_cat in cats:
|
||||||
|
|
||||||
if is_valid:
|
|
||||||
st.caption(f"Editing: **{target_cat}**")
|
st.caption(f"Editing: **{target_cat}**")
|
||||||
|
|
||||||
# RENAME
|
|
||||||
rename_val = st.text_input("Rename to:", value=target_cat, key=f"ren_{target_cat}")
|
rename_val = st.text_input("Rename to:", value=target_cat, key=f"ren_{target_cat}")
|
||||||
if st.button("💾 Save Name", key=f"save_{target_cat}", use_container_width=True):
|
if st.button("💾 Save", key=f"save_{target_cat}", use_container_width=True):
|
||||||
if rename_val and rename_val != target_cat:
|
if rename_val and rename_val != target_cat:
|
||||||
SorterEngine.rename_category(target_cat, rename_val)
|
SorterEngine.rename_category(target_cat, rename_val)
|
||||||
st.session_state.t5_active_cat = rename_val
|
st.session_state.t5_active_cat = rename_val
|
||||||
st.rerun()
|
st.rerun()
|
||||||
|
|
||||||
st.markdown("---")
|
st.markdown("---")
|
||||||
|
if st.button("🗑️ Delete", key=f"del_cat_{target_cat}", type="primary", use_container_width=True):
|
||||||
# DELETE
|
|
||||||
if st.button("🗑️ Delete Category", key=f"del_cat_{target_cat}", type="primary", use_container_width=True):
|
|
||||||
SorterEngine.delete_category(target_cat)
|
SorterEngine.delete_category(target_cat)
|
||||||
st.rerun()
|
st.rerun()
|
||||||
else:
|
else:
|
||||||
st.info("Select a valid category to edit.")
|
st.info("Select a valid category to edit.")
|
||||||
|
|
||||||
|
|
||||||
# ==========================================
|
# ... (Gallery Grid code remains exactly the same) ...
|
||||||
# 3. FRAGMENT: GALLERY GRID
|
|
||||||
# ==========================================
|
|
||||||
@st.fragment
|
@st.fragment
|
||||||
def render_gallery_grid(current_batch, quality, grid_cols):
|
def render_gallery_grid(current_batch, quality, grid_cols):
|
||||||
staged = SorterEngine.get_staged_data()
|
staged = SorterEngine.get_staged_data()
|
||||||
|
history = SorterEngine.get_processed_log()
|
||||||
selected_cat = st.session_state.get("t5_active_cat", "Default")
|
selected_cat = st.session_state.get("t5_active_cat", "Default")
|
||||||
tagging_disabled = selected_cat.startswith("---")
|
tagging_disabled = selected_cat.startswith("---")
|
||||||
|
|
||||||
|
# --- NEW: LOAD ALL IMAGES IN PARALLEL ---
|
||||||
|
# This runs multithreaded and is much faster than the old loop
|
||||||
|
batch_cache = SorterEngine.load_batch_parallel(current_batch, quality)
|
||||||
|
|
||||||
cols = st.columns(grid_cols)
|
cols = st.columns(grid_cols)
|
||||||
|
|
||||||
for idx, img_path in enumerate(current_batch):
|
for idx, img_path in enumerate(current_batch):
|
||||||
unique_key = f"frag_{os.path.basename(img_path)}"
|
unique_key = f"frag_{os.path.basename(img_path)}"
|
||||||
|
|
||||||
with cols[idx % grid_cols]:
|
with cols[idx % grid_cols]:
|
||||||
is_staged = img_path in staged
|
is_staged = img_path in staged
|
||||||
|
is_processed = img_path in history
|
||||||
|
|
||||||
with st.container(border=True):
|
with st.container(border=True):
|
||||||
# Header
|
|
||||||
c_head1, c_head2 = st.columns([5, 1])
|
c_head1, c_head2 = st.columns([5, 1])
|
||||||
c_head1.caption(os.path.basename(img_path)[:15])
|
c_head1.caption(os.path.basename(img_path)[:15])
|
||||||
|
c_head2.button("❌", key=f"del_{unique_key}", on_click=cb_delete_image, args=(img_path,))
|
||||||
# DELETE (Callback)
|
|
||||||
c_head2.button("❌", key=f"del_{unique_key}",
|
|
||||||
on_click=cb_delete_image, args=(img_path,))
|
|
||||||
|
|
||||||
# STATUS
|
|
||||||
if is_staged:
|
if is_staged:
|
||||||
st.success(f"🏷️ {staged[img_path]['cat']}")
|
st.success(f"🏷️ {staged[img_path]['cat']}")
|
||||||
|
elif is_processed:
|
||||||
# IMAGE
|
st.info(f"✅ {history[img_path]['action']} -> {history[img_path]['cat']}")
|
||||||
img_data = SorterEngine.compress_for_web(img_path, quality)
|
|
||||||
if img_data:
|
# --- CHANGED: USE PRE-LOADED DATA ---
|
||||||
|
img_data = batch_cache.get(img_path)
|
||||||
|
if img_data:
|
||||||
st.image(img_data, use_container_width=True)
|
st.image(img_data, use_container_width=True)
|
||||||
|
|
||||||
# ACTIONS (Callbacks)
|
|
||||||
if not is_staged:
|
if not is_staged:
|
||||||
st.button("Tag", key=f"tag_{unique_key}",
|
st.button("Tag", key=f"tag_{unique_key}", disabled=tagging_disabled, use_container_width=True,
|
||||||
disabled=tagging_disabled,
|
|
||||||
use_container_width=True,
|
|
||||||
on_click=cb_tag_image, args=(img_path, selected_cat))
|
on_click=cb_tag_image, args=(img_path, selected_cat))
|
||||||
else:
|
else:
|
||||||
st.button("Untag", key=f"untag_{unique_key}",
|
st.button("Untag", key=f"untag_{unique_key}", use_container_width=True,
|
||||||
use_container_width=True,
|
|
||||||
on_click=cb_untag_image, args=(img_path,))
|
on_click=cb_untag_image, args=(img_path,))
|
||||||
|
|
||||||
|
|
||||||
# ==========================================
|
# ... (Batch Actions code remains exactly the same) ...
|
||||||
# 4. FRAGMENT: BATCH ACTIONS
|
|
||||||
# ==========================================
|
|
||||||
@st.fragment
|
@st.fragment
|
||||||
def render_batch_actions(current_batch, path_o, page_num):
|
def render_batch_actions(current_batch, path_o, page_num, path_s):
|
||||||
st.write(f"### 🚀 Batch Actions (Page {page_num})")
|
st.write(f"### 🚀 Processing Actions")
|
||||||
|
st.caption("Settings apply to both Page and Global actions.")
|
||||||
|
c_set1, c_set2 = st.columns(2)
|
||||||
|
op_mode = c_set1.radio("Tagged Files:", ["Move", "Copy"], horizontal=True, key="t5_op_mode")
|
||||||
|
cleanup = c_set2.radio("Untagged Files:", ["Keep", "Move to Unused", "Delete"], horizontal=True, key="t5_cleanup_mode")
|
||||||
|
st.divider()
|
||||||
|
c_btn1, c_btn2 = st.columns(2)
|
||||||
|
|
||||||
# We use columns to organize the settings clearly
|
if c_btn1.button(f"APPLY PAGE {page_num}", type="secondary", use_container_width=True,
|
||||||
c_set1, c_set2, c_btn = st.columns([2, 2, 1.5], vertical_alignment="bottom")
|
on_click=cb_apply_batch, args=(current_batch, path_o, cleanup, op_mode)):
|
||||||
|
st.toast(f"Page {page_num} Applied!")
|
||||||
# 1. Operation for TAGGED files
|
st.rerun()
|
||||||
op_mode = c_set1.radio("Tagged Files:", ["Move", "Copy"],
|
|
||||||
horizontal=True, key="t5_op_mode")
|
if c_btn2.button("APPLY ALL (GLOBAL)", type="primary", use_container_width=True,
|
||||||
|
help="Process ALL tagged files across all pages.",
|
||||||
# 2. Action for UNTAGGED files
|
on_click=cb_apply_global, args=(path_o, cleanup, op_mode, path_s)):
|
||||||
cleanup = c_set2.radio("Untagged Files:", ["Keep", "Move to Unused", "Delete"],
|
st.toast("Global Apply Complete!")
|
||||||
horizontal=True, key="t5_cleanup_mode")
|
|
||||||
|
|
||||||
# 3. Apply Button
|
|
||||||
# Note: We added 'op_mode' to the args
|
|
||||||
if c_btn.button(f"APPLY ({op_mode})", type="primary", use_container_width=True,
|
|
||||||
on_click=cb_apply_batch, args=(current_batch, path_o, cleanup, op_mode)):
|
|
||||||
st.success("Batch processed!")
|
|
||||||
# Rerun to show changes (files disappearing or remaining depending on copy/move)
|
|
||||||
st.rerun()
|
st.rerun()
|
||||||
|
|
||||||
|
|
||||||
# ==========================================
|
# ==========================================
|
||||||
# 5. MAIN RENDERER
|
# 4. MAIN RENDERER
|
||||||
# ==========================================
|
# ==========================================
|
||||||
def render(quality, profile_name):
|
def render(quality, profile_name):
|
||||||
st.subheader("🖼️ Gallery Staging Sorter")
|
st.subheader("🖼️ Gallery Staging Sorter")
|
||||||
|
|
||||||
# Init State
|
# Init Mutation ID (This triggers the scanner cache refresh)
|
||||||
|
if 't5_file_id' not in st.session_state: st.session_state.t5_file_id = 0
|
||||||
if 't5_page' not in st.session_state: st.session_state.t5_page = 0
|
if 't5_page' not in st.session_state: st.session_state.t5_page = 0
|
||||||
|
|
||||||
# Load Paths
|
|
||||||
profiles = SorterEngine.load_profiles()
|
profiles = SorterEngine.load_profiles()
|
||||||
p_data = profiles.get(profile_name, {})
|
p_data = profiles.get(profile_name, {})
|
||||||
|
|
||||||
c1, c2 = st.columns(2)
|
c1, c2 = st.columns(2)
|
||||||
path_s = c1.text_input("Source Folder", value=p_data.get("tab5_source", "/storage"), key="t5_s")
|
path_s = c1.text_input("Source Folder", value=p_data.get("tab5_source", "/storage"), key="t5_s")
|
||||||
path_o = c2.text_input("Output Folder", value=p_data.get("tab5_out", "/storage"), key="t5_o")
|
path_o = c2.text_input("Output Folder", value=p_data.get("tab5_out", "/storage"), key="t5_o")
|
||||||
@@ -213,30 +197,30 @@ def render(quality, profile_name):
|
|||||||
if path_s != p_data.get("tab5_source") or path_o != p_data.get("tab5_out"):
|
if path_s != p_data.get("tab5_source") or path_o != p_data.get("tab5_out"):
|
||||||
if st.button("💾 Save Settings"):
|
if st.button("💾 Save Settings"):
|
||||||
SorterEngine.save_tab_paths(profile_name, t5_s=path_s, t5_o=path_o)
|
SorterEngine.save_tab_paths(profile_name, t5_s=path_s, t5_o=path_o)
|
||||||
|
# Saving settings might mean new folder, so we trigger refresh
|
||||||
|
trigger_refresh()
|
||||||
st.rerun()
|
st.rerun()
|
||||||
|
|
||||||
if not os.path.exists(path_s): return
|
if not os.path.exists(path_s): return
|
||||||
|
|
||||||
# --- RENDER SIDEBAR ---
|
|
||||||
with st.sidebar:
|
with st.sidebar:
|
||||||
render_sidebar_content()
|
render_sidebar_content()
|
||||||
|
|
||||||
# --- VIEW SETTINGS ---
|
|
||||||
with st.expander("👀 View Settings"):
|
with st.expander("👀 View Settings"):
|
||||||
c_v1, c_v2 = st.columns(2)
|
c_v1, c_v2 = st.columns(2)
|
||||||
page_size = c_v1.slider("Images per Page", 12, 100, 24, 4)
|
page_size = c_v1.slider("Images per Page", 12, 100, 24, 4)
|
||||||
grid_cols = c_v2.slider("Grid Columns", 2, 8, 4)
|
grid_cols = c_v2.slider("Grid Columns", 2, 8, 4)
|
||||||
|
|
||||||
# --- DATA & MATH ---
|
# --- USING CACHED LOADER ---
|
||||||
all_images = SorterEngine.get_images(path_s, recursive=True)
|
# We pass the mutation ID. If ID is same as last run, scan is SKIPPED.
|
||||||
|
all_images = get_cached_images(path_s, st.session_state.t5_file_id)
|
||||||
|
|
||||||
if not all_images:
|
if not all_images:
|
||||||
st.info("No images found.")
|
st.info("No images found.")
|
||||||
return
|
return
|
||||||
|
|
||||||
total_items = len(all_images)
|
total_items = len(all_images)
|
||||||
total_pages = math.ceil(total_items / page_size)
|
total_pages = math.ceil(total_items / page_size)
|
||||||
|
|
||||||
# Safety Bounds Check
|
|
||||||
if st.session_state.t5_page >= total_pages: st.session_state.t5_page = max(0, total_pages - 1)
|
if st.session_state.t5_page >= total_pages: st.session_state.t5_page = max(0, total_pages - 1)
|
||||||
if st.session_state.t5_page < 0: st.session_state.t5_page = 0
|
if st.session_state.t5_page < 0: st.session_state.t5_page = 0
|
||||||
|
|
||||||
@@ -244,46 +228,18 @@ def render(quality, profile_name):
|
|||||||
end_idx = start_idx + page_size
|
end_idx = start_idx + page_size
|
||||||
current_batch = all_images[start_idx:end_idx]
|
current_batch = all_images[start_idx:end_idx]
|
||||||
|
|
||||||
# --- NAVIGATION BAR COMPONENT ---
|
|
||||||
def nav_controls(key_suffix):
|
def nav_controls(key_suffix):
|
||||||
# New Layout: [Prev] [Input Box] ["/ 15"] [Next]
|
|
||||||
c1, c2, c3, c4 = st.columns([1.5, 1, 0.5, 1.5], vertical_alignment="center")
|
c1, c2, c3, c4 = st.columns([1.5, 1, 0.5, 1.5], vertical_alignment="center")
|
||||||
|
c1.button("⬅️ Prev", disabled=(st.session_state.t5_page == 0), on_click=cb_change_page, args=(-1,), key=f"p_{key_suffix}", use_container_width=True)
|
||||||
# 1. Previous Button
|
c2.number_input("Page", min_value=1, max_value=total_pages, value=st.session_state.t5_page + 1, step=1, label_visibility="collapsed", key=f"jump_{key_suffix}", on_change=cb_jump_page, args=(f"jump_{key_suffix}",))
|
||||||
c1.button("⬅️ Prev",
|
|
||||||
disabled=(st.session_state.t5_page == 0),
|
|
||||||
on_click=cb_change_page, args=(-1,),
|
|
||||||
key=f"p_{key_suffix}", use_container_width=True)
|
|
||||||
|
|
||||||
# 2. Page Selector (Number Input)
|
|
||||||
c2.number_input(
|
|
||||||
"Page",
|
|
||||||
min_value=1, max_value=total_pages,
|
|
||||||
value=st.session_state.t5_page + 1,
|
|
||||||
step=1,
|
|
||||||
label_visibility="collapsed",
|
|
||||||
key=f"jump_{key_suffix}",
|
|
||||||
on_change=cb_jump_page, args=(f"jump_{key_suffix}",)
|
|
||||||
)
|
|
||||||
|
|
||||||
# 3. Total Page Count Display
|
|
||||||
c3.markdown(f"<div style='text-align: left; font-weight: bold;'>/ {total_pages}</div>", unsafe_allow_html=True)
|
c3.markdown(f"<div style='text-align: left; font-weight: bold;'>/ {total_pages}</div>", unsafe_allow_html=True)
|
||||||
|
c4.button("Next ➡️", disabled=(st.session_state.t5_page >= total_pages - 1), on_click=cb_change_page, args=(1,), key=f"n_{key_suffix}", use_container_width=True)
|
||||||
# 4. Next Button
|
|
||||||
c4.button("Next ➡️",
|
|
||||||
disabled=(st.session_state.t5_page >= total_pages - 1),
|
|
||||||
on_click=cb_change_page, args=(1,),
|
|
||||||
key=f"n_{key_suffix}", use_container_width=True)
|
|
||||||
|
|
||||||
# --- RENDER PAGE ---
|
|
||||||
st.divider()
|
st.divider()
|
||||||
nav_controls("top") # Top Nav
|
nav_controls("top")
|
||||||
|
render_gallery_grid(current_batch, quality, grid_cols)
|
||||||
|
st.divider()
|
||||||
|
nav_controls("bottom")
|
||||||
|
st.divider()
|
||||||
|
|
||||||
render_gallery_grid(current_batch, quality, grid_cols) # Grid
|
render_batch_actions(current_batch, path_o, st.session_state.t5_page + 1, path_s)
|
||||||
|
|
||||||
st.divider()
|
|
||||||
nav_controls("bottom") # Bottom Nav
|
|
||||||
st.divider()
|
|
||||||
|
|
||||||
# Batch Actions
|
|
||||||
render_batch_actions(current_batch, path_o, st.session_state.t5_page + 1)
|
|
||||||
Reference in New Issue
Block a user