global-button #1
128
engine.py
128
engine.py
@@ -10,30 +10,24 @@ class SorterEngine:
|
||||
# --- 1. DATABASE INITIALIZATION ---
|
||||
@staticmethod
|
||||
def init_db():
|
||||
"""Initializes all SQLite tables for the multi-tab system."""
|
||||
"""Initializes tables, including the new HISTORY log."""
|
||||
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Profiles Table: 9 columns for independent tab paths
|
||||
# Existing tables...
|
||||
cursor.execute('''CREATE TABLE IF NOT EXISTS profiles
|
||||
(name TEXT PRIMARY KEY,
|
||||
tab1_target TEXT,
|
||||
tab2_target TEXT, tab2_control TEXT,
|
||||
tab4_source TEXT, tab4_out TEXT,
|
||||
mode TEXT,
|
||||
tab5_source TEXT, tab5_out TEXT)''')
|
||||
|
||||
(name TEXT PRIMARY KEY, tab1_target TEXT, tab2_target TEXT, tab2_control TEXT,
|
||||
tab4_source TEXT, tab4_out TEXT, mode TEXT, tab5_source TEXT, tab5_out TEXT)''')
|
||||
cursor.execute('''CREATE TABLE IF NOT EXISTS folder_ids (path TEXT PRIMARY KEY, folder_id INTEGER)''')
|
||||
cursor.execute('''CREATE TABLE IF NOT EXISTS categories (name TEXT PRIMARY KEY)''')
|
||||
|
||||
# Staging Area: Tracks pending renames for the Gallery Tab
|
||||
cursor.execute('''CREATE TABLE IF NOT EXISTS staging_area
|
||||
(original_path TEXT PRIMARY KEY,
|
||||
target_category TEXT,
|
||||
new_name TEXT,
|
||||
is_marked INTEGER DEFAULT 0)''')
|
||||
(original_path TEXT PRIMARY KEY, target_category TEXT, new_name TEXT, is_marked INTEGER DEFAULT 0)''')
|
||||
|
||||
# --- NEW: HISTORY TABLE ---
|
||||
cursor.execute('''CREATE TABLE IF NOT EXISTS processed_log
|
||||
(source_path TEXT PRIMARY KEY, category TEXT, action_type TEXT)''')
|
||||
|
||||
# Seed default categories
|
||||
# Seed categories if empty
|
||||
cursor.execute("SELECT COUNT(*) FROM categories")
|
||||
if cursor.fetchone()[0] == 0:
|
||||
for cat in ["_TRASH", "Default", "Action", "Solo"]:
|
||||
@@ -69,6 +63,31 @@ class SorterEngine:
|
||||
cursor.execute("INSERT OR REPLACE INTO profiles VALUES (?, ?, ?, ?, ?, ?, ?, ?, ?)", new_values)
|
||||
conn.commit()
|
||||
conn.close()
|
||||
@staticmethod
|
||||
def load_batch_parallel(image_paths, quality):
|
||||
"""
|
||||
Multithreaded loader: Compresses multiple images in parallel.
|
||||
Returns a dictionary {path: bytes_io}
|
||||
"""
|
||||
import concurrent.futures
|
||||
|
||||
results = {}
|
||||
|
||||
# Helper function to run in thread
|
||||
def process_one(path):
|
||||
return path, SorterEngine.compress_for_web(path, quality)
|
||||
|
||||
# Use ThreadPool to parallelize IO-bound tasks
|
||||
with concurrent.futures.ThreadPoolExecutor(max_workers=8) as executor:
|
||||
# Submit all tasks
|
||||
future_to_path = {executor.submit(process_one, p): p for p in image_paths}
|
||||
|
||||
# Gather results as they complete
|
||||
for future in concurrent.futures.as_completed(future_to_path):
|
||||
path, data = future.result()
|
||||
results[path] = data
|
||||
|
||||
return results
|
||||
|
||||
@staticmethod
|
||||
def load_profiles():
|
||||
@@ -234,22 +253,19 @@ class SorterEngine:
|
||||
return {r[0]: {"cat": r[1], "name": r[2], "marked": r[3]} for r in rows}
|
||||
|
||||
@staticmethod
|
||||
def commit_staging(output_root, cleanup_mode, source_root=None):
|
||||
"""Global commit directly to output root (No Subfolders)."""
|
||||
def commit_global(output_root, cleanup_mode, operation="Move", source_root=None):
|
||||
"""Commits ALL staged files (Global Apply)."""
|
||||
data = SorterEngine.get_staged_data()
|
||||
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
||||
cursor = conn.cursor()
|
||||
staged_paths = set(data.keys())
|
||||
|
||||
if not os.path.exists(output_root):
|
||||
os.makedirs(output_root, exist_ok=True)
|
||||
if not os.path.exists(output_root): os.makedirs(output_root, exist_ok=True)
|
||||
|
||||
# 1. Process all Staged Items
|
||||
for old_p, info in data.items():
|
||||
if os.path.exists(old_p):
|
||||
# CHANGED: Direct move to root
|
||||
final_dst = os.path.join(output_root, info['name'])
|
||||
|
||||
# Collision Safety for global commit
|
||||
if os.path.exists(final_dst):
|
||||
root, ext = os.path.splitext(info['name'])
|
||||
c = 1
|
||||
@@ -257,17 +273,28 @@ class SorterEngine:
|
||||
final_dst = os.path.join(output_root, f"{root}_{c}{ext}")
|
||||
c += 1
|
||||
|
||||
shutil.move(old_p, final_dst)
|
||||
|
||||
if operation == "Copy":
|
||||
shutil.copy2(old_p, final_dst)
|
||||
else:
|
||||
shutil.move(old_p, final_dst)
|
||||
|
||||
# Log History
|
||||
cursor.execute("INSERT OR REPLACE INTO processed_log VALUES (?, ?, ?)",
|
||||
(old_p, info['cat'], operation))
|
||||
|
||||
# 2. Global Cleanup (Optional)
|
||||
# Only run cleanup if explicitly asked, as global cleanup is risky
|
||||
if cleanup_mode != "Keep" and source_root:
|
||||
for img_p in SorterEngine.get_images(source_root, recursive=True):
|
||||
if img_p not in staged_paths:
|
||||
all_imgs = SorterEngine.get_images(source_root, recursive=True)
|
||||
for img_p in all_imgs:
|
||||
if img_p not in data: # Not currently staged
|
||||
if cleanup_mode == "Move to Unused":
|
||||
un_dir = os.path.join(source_root, "unused")
|
||||
os.makedirs(un_dir, exist_ok=True)
|
||||
shutil.move(img_p, os.path.join(un_dir, os.path.basename(img_p)))
|
||||
elif cleanup_mode == "Delete": os.remove(img_p)
|
||||
|
||||
unused_dir = os.path.join(source_root, "unused")
|
||||
os.makedirs(unused_dir, exist_ok=True)
|
||||
shutil.move(img_p, os.path.join(unused_dir, os.path.basename(img_p)))
|
||||
elif cleanup_mode == "Delete":
|
||||
os.remove(img_p)
|
||||
|
||||
cursor.execute("DELETE FROM staging_area")
|
||||
conn.commit()
|
||||
conn.close()
|
||||
@@ -346,28 +373,34 @@ class SorterEngine:
|
||||
if 'c_dst' in action and os.path.exists(action['c_dst']):
|
||||
shutil.move(action['c_dst'], action['c_src'])
|
||||
|
||||
@staticmethod
|
||||
def get_processed_log():
|
||||
"""Retrieves history of processed files."""
|
||||
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
||||
cursor = conn.cursor()
|
||||
cursor.execute("SELECT * FROM processed_log")
|
||||
rows = cursor.fetchall()
|
||||
conn.close()
|
||||
return {r[0]: {"cat": r[1], "action": r[2]} for r in rows}
|
||||
|
||||
@staticmethod
|
||||
def commit_batch(file_list, output_root, cleanup_mode, operation="Move"):
|
||||
"""
|
||||
Commits files with support for both MOVE and COPY.
|
||||
"""
|
||||
"""Commits specified files and LOGS them to history."""
|
||||
data = SorterEngine.get_staged_data()
|
||||
conn = sqlite3.connect(SorterEngine.DB_PATH)
|
||||
cursor = conn.cursor()
|
||||
|
||||
# Ensure output root exists
|
||||
if not os.path.exists(output_root):
|
||||
os.makedirs(output_root, exist_ok=True)
|
||||
if not os.path.exists(output_root): os.makedirs(output_root, exist_ok=True)
|
||||
|
||||
for file_path in file_list:
|
||||
if not os.path.exists(file_path): continue
|
||||
|
||||
# --- CASE A: File is TAGGED ---
|
||||
# --- CASE A: Tagged ---
|
||||
if file_path in data and data[file_path]['marked']:
|
||||
info = data[file_path]
|
||||
final_dst = os.path.join(output_root, info['name'])
|
||||
|
||||
# Collision Safety
|
||||
# Collision Check
|
||||
if os.path.exists(final_dst):
|
||||
root, ext = os.path.splitext(info['name'])
|
||||
c = 1
|
||||
@@ -375,22 +408,21 @@ class SorterEngine:
|
||||
final_dst = os.path.join(output_root, f"{root}_{c}{ext}")
|
||||
c += 1
|
||||
|
||||
# OPERATION CHECK: Move vs Copy
|
||||
# Action
|
||||
if operation == "Copy":
|
||||
shutil.copy2(file_path, final_dst) # copy2 preserves metadata
|
||||
shutil.copy2(file_path, final_dst)
|
||||
else:
|
||||
shutil.move(file_path, final_dst)
|
||||
|
||||
# Remove from staging database
|
||||
# Update DB: Remove from Staging, Add to History
|
||||
cursor.execute("DELETE FROM staging_area WHERE original_path = ?", (file_path,))
|
||||
cursor.execute("INSERT OR REPLACE INTO processed_log VALUES (?, ?, ?)",
|
||||
(file_path, info['cat'], operation))
|
||||
|
||||
# --- CASE B: File is UNTAGGED (Cleanup) ---
|
||||
# Note: If we COPIED a tagged file, the original stays in source,
|
||||
# but this 'elif' ensures we don't accidentally delete it as 'untagged'.
|
||||
# --- CASE B: Cleanup ---
|
||||
elif cleanup_mode != "Keep":
|
||||
if cleanup_mode == "Move to Unused":
|
||||
parent = os.path.dirname(file_path)
|
||||
unused_dir = os.path.join(parent, "unused")
|
||||
unused_dir = os.path.join(os.path.dirname(file_path), "unused")
|
||||
os.makedirs(unused_dir, exist_ok=True)
|
||||
shutil.move(file_path, os.path.join(unused_dir, os.path.basename(file_path)))
|
||||
elif cleanup_mode == "Delete":
|
||||
|
||||
@@ -4,57 +4,69 @@ import math
|
||||
from engine import SorterEngine
|
||||
|
||||
# ==========================================
|
||||
# 1. GLOBAL CALLBACKS (Prevents Page Refresh)
|
||||
# 1. CACHED DATA LOADER (The Fix)
|
||||
# ==========================================
|
||||
@st.cache_data(show_spinner=False)
|
||||
def get_cached_images(path, mutation_id):
|
||||
"""
|
||||
Scans the folder ONLY when 'path' or 'mutation_id' changes.
|
||||
Navigating pages does NOT change these, so it remains instant.
|
||||
"""
|
||||
return SorterEngine.get_images(path, recursive=True)
|
||||
|
||||
|
||||
# ==========================================
|
||||
# 2. CALLBACKS (Updated with Refresh Logic)
|
||||
# ==========================================
|
||||
def trigger_refresh():
|
||||
"""Increments the mutation counter to force a file re-scan."""
|
||||
if 't5_file_id' not in st.session_state: st.session_state.t5_file_id = 0
|
||||
st.session_state.t5_file_id += 1
|
||||
|
||||
def cb_tag_image(img_path, selected_cat):
|
||||
"""Tags an image. Updates DB immediately."""
|
||||
if selected_cat.startswith("---") or selected_cat == "":
|
||||
st.toast("⚠️ Select a valid category first!", icon="🚫")
|
||||
return
|
||||
|
||||
staged = SorterEngine.get_staged_data()
|
||||
ext = os.path.splitext(img_path)[1]
|
||||
|
||||
# Auto-increment logic
|
||||
count = len([v for v in staged.values() if v['cat'] == selected_cat]) + 1
|
||||
new_name = f"{selected_cat}_{count:03d}{ext}"
|
||||
|
||||
SorterEngine.stage_image(img_path, selected_cat, new_name)
|
||||
# Note: Tagging does NOT need a file re-scan, just a grid refresh.
|
||||
|
||||
def cb_untag_image(img_path):
|
||||
"""Untags an image."""
|
||||
SorterEngine.clear_staged_item(img_path)
|
||||
|
||||
def cb_delete_image(img_path):
|
||||
"""Moves image to trash."""
|
||||
SorterEngine.delete_to_trash(img_path)
|
||||
trigger_refresh() # Force re-scan so the image disappears from the list
|
||||
|
||||
def cb_apply_batch(current_batch, path_o, cleanup_mode, operation):
|
||||
"""Commits the batch with the specified operation (Move/Copy)."""
|
||||
SorterEngine.commit_batch(current_batch, path_o, cleanup_mode, operation)
|
||||
trigger_refresh() # Force re-scan to remove moved files
|
||||
|
||||
def cb_apply_global(path_o, cleanup_mode, operation, path_s):
|
||||
SorterEngine.commit_global(path_o, cleanup_mode, operation, source_root=path_s)
|
||||
trigger_refresh() # Force re-scan
|
||||
|
||||
def cb_change_page(delta):
|
||||
"""Updates page number (-1 or +1)."""
|
||||
if 't5_page' not in st.session_state:
|
||||
st.session_state.t5_page = 0
|
||||
if 't5_page' not in st.session_state: st.session_state.t5_page = 0
|
||||
st.session_state.t5_page += delta
|
||||
# No trigger_refresh() here -> This is why page turning is now instant!
|
||||
|
||||
def cb_jump_page(k):
|
||||
"""Updates page number from direct input box."""
|
||||
val = st.session_state[k]
|
||||
st.session_state.t5_page = val - 1
|
||||
|
||||
|
||||
# ==========================================
|
||||
# 2. FRAGMENT: SIDEBAR (Category Manager)
|
||||
# 3. FRAGMENTS (Sidebar, Grid, Batch)
|
||||
# ==========================================
|
||||
# ... (Sidebar code remains exactly the same) ...
|
||||
@st.fragment
|
||||
def render_sidebar_content():
|
||||
st.divider()
|
||||
st.subheader("🏷️ Category Manager")
|
||||
|
||||
# --- PREPARE LIST (With Separators) ---
|
||||
cats = SorterEngine.get_categories()
|
||||
processed_cats = []
|
||||
last_char = ""
|
||||
@@ -66,26 +78,16 @@ def render_sidebar_content():
|
||||
processed_cats.append(cat)
|
||||
last_char = current_char
|
||||
|
||||
# --- STATE SYNC ---
|
||||
if "t5_active_cat" not in st.session_state:
|
||||
st.session_state.t5_active_cat = cats[0] if cats else "Default"
|
||||
|
||||
# Fallback if selection was deleted
|
||||
if "t5_active_cat" not in st.session_state: st.session_state.t5_active_cat = cats[0] if cats else "Default"
|
||||
current_selection = st.session_state.t5_active_cat
|
||||
if not current_selection.startswith("---") and current_selection not in cats:
|
||||
st.session_state.t5_active_cat = cats[0] if cats else "Default"
|
||||
|
||||
# --- RADIO SELECTION ---
|
||||
selection = st.radio("Active Tag", processed_cats, key="t5_radio_select")
|
||||
|
||||
if not selection.startswith("---"):
|
||||
st.session_state.t5_active_cat = selection
|
||||
if not selection.startswith("---"): st.session_state.t5_active_cat = selection
|
||||
|
||||
st.divider()
|
||||
|
||||
# --- TABS: ADD / EDIT ---
|
||||
tab_add, tab_edit = st.tabs(["➕ Add", "✏️ Edit"])
|
||||
|
||||
with tab_add:
|
||||
c1, c2 = st.columns([3, 1])
|
||||
new_cat = c1.text_input("New Name", label_visibility="collapsed", placeholder="New...", key="t5_new_cat")
|
||||
@@ -93,119 +95,101 @@ def render_sidebar_content():
|
||||
if new_cat:
|
||||
SorterEngine.add_category(new_cat)
|
||||
st.rerun()
|
||||
|
||||
with tab_edit:
|
||||
target_cat = st.session_state.t5_active_cat
|
||||
is_valid = target_cat and not target_cat.startswith("---") and target_cat in cats
|
||||
|
||||
if is_valid:
|
||||
if target_cat and not target_cat.startswith("---") and target_cat in cats:
|
||||
st.caption(f"Editing: **{target_cat}**")
|
||||
|
||||
# RENAME
|
||||
rename_val = st.text_input("Rename to:", value=target_cat, key=f"ren_{target_cat}")
|
||||
if st.button("💾 Save Name", key=f"save_{target_cat}", use_container_width=True):
|
||||
if st.button("💾 Save", key=f"save_{target_cat}", use_container_width=True):
|
||||
if rename_val and rename_val != target_cat:
|
||||
SorterEngine.rename_category(target_cat, rename_val)
|
||||
st.session_state.t5_active_cat = rename_val
|
||||
st.rerun()
|
||||
|
||||
st.markdown("---")
|
||||
|
||||
# DELETE
|
||||
if st.button("🗑️ Delete Category", key=f"del_cat_{target_cat}", type="primary", use_container_width=True):
|
||||
if st.button("🗑️ Delete", key=f"del_cat_{target_cat}", type="primary", use_container_width=True):
|
||||
SorterEngine.delete_category(target_cat)
|
||||
st.rerun()
|
||||
else:
|
||||
st.info("Select a valid category to edit.")
|
||||
|
||||
|
||||
# ==========================================
|
||||
# 3. FRAGMENT: GALLERY GRID
|
||||
# ==========================================
|
||||
# ... (Gallery Grid code remains exactly the same) ...
|
||||
@st.fragment
|
||||
def render_gallery_grid(current_batch, quality, grid_cols):
|
||||
staged = SorterEngine.get_staged_data()
|
||||
history = SorterEngine.get_processed_log()
|
||||
selected_cat = st.session_state.get("t5_active_cat", "Default")
|
||||
tagging_disabled = selected_cat.startswith("---")
|
||||
|
||||
# --- NEW: LOAD ALL IMAGES IN PARALLEL ---
|
||||
# This runs multithreaded and is much faster than the old loop
|
||||
batch_cache = SorterEngine.load_batch_parallel(current_batch, quality)
|
||||
|
||||
cols = st.columns(grid_cols)
|
||||
|
||||
for idx, img_path in enumerate(current_batch):
|
||||
unique_key = f"frag_{os.path.basename(img_path)}"
|
||||
|
||||
with cols[idx % grid_cols]:
|
||||
is_staged = img_path in staged
|
||||
is_processed = img_path in history
|
||||
|
||||
with st.container(border=True):
|
||||
# Header
|
||||
c_head1, c_head2 = st.columns([5, 1])
|
||||
c_head1.caption(os.path.basename(img_path)[:15])
|
||||
|
||||
# DELETE (Callback)
|
||||
c_head2.button("❌", key=f"del_{unique_key}",
|
||||
on_click=cb_delete_image, args=(img_path,))
|
||||
c_head2.button("❌", key=f"del_{unique_key}", on_click=cb_delete_image, args=(img_path,))
|
||||
|
||||
# STATUS
|
||||
if is_staged:
|
||||
st.success(f"🏷️ {staged[img_path]['cat']}")
|
||||
|
||||
# IMAGE
|
||||
img_data = SorterEngine.compress_for_web(img_path, quality)
|
||||
if img_data:
|
||||
elif is_processed:
|
||||
st.info(f"✅ {history[img_path]['action']} -> {history[img_path]['cat']}")
|
||||
|
||||
# --- CHANGED: USE PRE-LOADED DATA ---
|
||||
img_data = batch_cache.get(img_path)
|
||||
if img_data:
|
||||
st.image(img_data, use_container_width=True)
|
||||
|
||||
# ACTIONS (Callbacks)
|
||||
if not is_staged:
|
||||
st.button("Tag", key=f"tag_{unique_key}",
|
||||
disabled=tagging_disabled,
|
||||
use_container_width=True,
|
||||
st.button("Tag", key=f"tag_{unique_key}", disabled=tagging_disabled, use_container_width=True,
|
||||
on_click=cb_tag_image, args=(img_path, selected_cat))
|
||||
else:
|
||||
st.button("Untag", key=f"untag_{unique_key}",
|
||||
use_container_width=True,
|
||||
st.button("Untag", key=f"untag_{unique_key}", use_container_width=True,
|
||||
on_click=cb_untag_image, args=(img_path,))
|
||||
|
||||
|
||||
# ==========================================
|
||||
# 4. FRAGMENT: BATCH ACTIONS
|
||||
# ==========================================
|
||||
# ... (Batch Actions code remains exactly the same) ...
|
||||
@st.fragment
|
||||
def render_batch_actions(current_batch, path_o, page_num):
|
||||
st.write(f"### 🚀 Batch Actions (Page {page_num})")
|
||||
def render_batch_actions(current_batch, path_o, page_num, path_s):
|
||||
st.write(f"### 🚀 Processing Actions")
|
||||
st.caption("Settings apply to both Page and Global actions.")
|
||||
c_set1, c_set2 = st.columns(2)
|
||||
op_mode = c_set1.radio("Tagged Files:", ["Move", "Copy"], horizontal=True, key="t5_op_mode")
|
||||
cleanup = c_set2.radio("Untagged Files:", ["Keep", "Move to Unused", "Delete"], horizontal=True, key="t5_cleanup_mode")
|
||||
st.divider()
|
||||
c_btn1, c_btn2 = st.columns(2)
|
||||
|
||||
# We use columns to organize the settings clearly
|
||||
c_set1, c_set2, c_btn = st.columns([2, 2, 1.5], vertical_alignment="bottom")
|
||||
|
||||
# 1. Operation for TAGGED files
|
||||
op_mode = c_set1.radio("Tagged Files:", ["Move", "Copy"],
|
||||
horizontal=True, key="t5_op_mode")
|
||||
|
||||
# 2. Action for UNTAGGED files
|
||||
cleanup = c_set2.radio("Untagged Files:", ["Keep", "Move to Unused", "Delete"],
|
||||
horizontal=True, key="t5_cleanup_mode")
|
||||
|
||||
# 3. Apply Button
|
||||
# Note: We added 'op_mode' to the args
|
||||
if c_btn.button(f"APPLY ({op_mode})", type="primary", use_container_width=True,
|
||||
on_click=cb_apply_batch, args=(current_batch, path_o, cleanup, op_mode)):
|
||||
st.success("Batch processed!")
|
||||
# Rerun to show changes (files disappearing or remaining depending on copy/move)
|
||||
if c_btn1.button(f"APPLY PAGE {page_num}", type="secondary", use_container_width=True,
|
||||
on_click=cb_apply_batch, args=(current_batch, path_o, cleanup, op_mode)):
|
||||
st.toast(f"Page {page_num} Applied!")
|
||||
st.rerun()
|
||||
|
||||
if c_btn2.button("APPLY ALL (GLOBAL)", type="primary", use_container_width=True,
|
||||
help="Process ALL tagged files across all pages.",
|
||||
on_click=cb_apply_global, args=(path_o, cleanup, op_mode, path_s)):
|
||||
st.toast("Global Apply Complete!")
|
||||
st.rerun()
|
||||
|
||||
|
||||
# ==========================================
|
||||
# 5. MAIN RENDERER
|
||||
# 4. MAIN RENDERER
|
||||
# ==========================================
|
||||
def render(quality, profile_name):
|
||||
st.subheader("🖼️ Gallery Staging Sorter")
|
||||
|
||||
# Init State
|
||||
# Init Mutation ID (This triggers the scanner cache refresh)
|
||||
if 't5_file_id' not in st.session_state: st.session_state.t5_file_id = 0
|
||||
if 't5_page' not in st.session_state: st.session_state.t5_page = 0
|
||||
|
||||
# Load Paths
|
||||
profiles = SorterEngine.load_profiles()
|
||||
p_data = profiles.get(profile_name, {})
|
||||
|
||||
c1, c2 = st.columns(2)
|
||||
path_s = c1.text_input("Source Folder", value=p_data.get("tab5_source", "/storage"), key="t5_s")
|
||||
path_o = c2.text_input("Output Folder", value=p_data.get("tab5_out", "/storage"), key="t5_o")
|
||||
@@ -213,30 +197,30 @@ def render(quality, profile_name):
|
||||
if path_s != p_data.get("tab5_source") or path_o != p_data.get("tab5_out"):
|
||||
if st.button("💾 Save Settings"):
|
||||
SorterEngine.save_tab_paths(profile_name, t5_s=path_s, t5_o=path_o)
|
||||
# Saving settings might mean new folder, so we trigger refresh
|
||||
trigger_refresh()
|
||||
st.rerun()
|
||||
|
||||
if not os.path.exists(path_s): return
|
||||
|
||||
# --- RENDER SIDEBAR ---
|
||||
with st.sidebar:
|
||||
render_sidebar_content()
|
||||
|
||||
# --- VIEW SETTINGS ---
|
||||
with st.expander("👀 View Settings"):
|
||||
c_v1, c_v2 = st.columns(2)
|
||||
page_size = c_v1.slider("Images per Page", 12, 100, 24, 4)
|
||||
grid_cols = c_v2.slider("Grid Columns", 2, 8, 4)
|
||||
|
||||
# --- DATA & MATH ---
|
||||
all_images = SorterEngine.get_images(path_s, recursive=True)
|
||||
# --- USING CACHED LOADER ---
|
||||
# We pass the mutation ID. If ID is same as last run, scan is SKIPPED.
|
||||
all_images = get_cached_images(path_s, st.session_state.t5_file_id)
|
||||
|
||||
if not all_images:
|
||||
st.info("No images found.")
|
||||
return
|
||||
|
||||
total_items = len(all_images)
|
||||
total_pages = math.ceil(total_items / page_size)
|
||||
|
||||
# Safety Bounds Check
|
||||
if st.session_state.t5_page >= total_pages: st.session_state.t5_page = max(0, total_pages - 1)
|
||||
if st.session_state.t5_page < 0: st.session_state.t5_page = 0
|
||||
|
||||
@@ -244,46 +228,18 @@ def render(quality, profile_name):
|
||||
end_idx = start_idx + page_size
|
||||
current_batch = all_images[start_idx:end_idx]
|
||||
|
||||
# --- NAVIGATION BAR COMPONENT ---
|
||||
def nav_controls(key_suffix):
|
||||
# New Layout: [Prev] [Input Box] ["/ 15"] [Next]
|
||||
c1, c2, c3, c4 = st.columns([1.5, 1, 0.5, 1.5], vertical_alignment="center")
|
||||
|
||||
# 1. Previous Button
|
||||
c1.button("⬅️ Prev",
|
||||
disabled=(st.session_state.t5_page == 0),
|
||||
on_click=cb_change_page, args=(-1,),
|
||||
key=f"p_{key_suffix}", use_container_width=True)
|
||||
|
||||
# 2. Page Selector (Number Input)
|
||||
c2.number_input(
|
||||
"Page",
|
||||
min_value=1, max_value=total_pages,
|
||||
value=st.session_state.t5_page + 1,
|
||||
step=1,
|
||||
label_visibility="collapsed",
|
||||
key=f"jump_{key_suffix}",
|
||||
on_change=cb_jump_page, args=(f"jump_{key_suffix}",)
|
||||
)
|
||||
|
||||
# 3. Total Page Count Display
|
||||
c1.button("⬅️ Prev", disabled=(st.session_state.t5_page == 0), on_click=cb_change_page, args=(-1,), key=f"p_{key_suffix}", use_container_width=True)
|
||||
c2.number_input("Page", min_value=1, max_value=total_pages, value=st.session_state.t5_page + 1, step=1, label_visibility="collapsed", key=f"jump_{key_suffix}", on_change=cb_jump_page, args=(f"jump_{key_suffix}",))
|
||||
c3.markdown(f"<div style='text-align: left; font-weight: bold;'>/ {total_pages}</div>", unsafe_allow_html=True)
|
||||
|
||||
# 4. Next Button
|
||||
c4.button("Next ➡️",
|
||||
disabled=(st.session_state.t5_page >= total_pages - 1),
|
||||
on_click=cb_change_page, args=(1,),
|
||||
key=f"n_{key_suffix}", use_container_width=True)
|
||||
c4.button("Next ➡️", disabled=(st.session_state.t5_page >= total_pages - 1), on_click=cb_change_page, args=(1,), key=f"n_{key_suffix}", use_container_width=True)
|
||||
|
||||
# --- RENDER PAGE ---
|
||||
st.divider()
|
||||
nav_controls("top") # Top Nav
|
||||
nav_controls("top")
|
||||
render_gallery_grid(current_batch, quality, grid_cols)
|
||||
st.divider()
|
||||
nav_controls("bottom")
|
||||
st.divider()
|
||||
|
||||
render_gallery_grid(current_batch, quality, grid_cols) # Grid
|
||||
|
||||
st.divider()
|
||||
nav_controls("bottom") # Bottom Nav
|
||||
st.divider()
|
||||
|
||||
# Batch Actions
|
||||
render_batch_actions(current_batch, path_o, st.session_state.t5_page + 1)
|
||||
render_batch_actions(current_batch, path_o, st.session_state.t5_page + 1, path_s)
|
||||
Reference in New Issue
Block a user