mirror of
https://github.com/comfyanonymous/ComfyUI.git
synced 2025-12-21 03:50:50 +08:00
refactor(2)/fix: skip double checking the existing files during fast check
This commit is contained in:
parent
a336c7c165
commit
31ec744317
@ -98,14 +98,19 @@ async def schedule_scans(roots: list[schemas_in.RootType]) -> schemas_out.AssetS
|
||||
|
||||
async def sync_seed_assets(roots: list[schemas_in.RootType]) -> None:
|
||||
t_total = time.perf_counter()
|
||||
created = 0
|
||||
skipped_existing = 0
|
||||
paths: list[str] = []
|
||||
try:
|
||||
existing_paths: set[str] = set()
|
||||
for r in roots:
|
||||
try:
|
||||
await _fast_db_consistency_pass(r)
|
||||
survivors = await _fast_db_consistency_pass(r, collect_existing_paths=True)
|
||||
if survivors:
|
||||
existing_paths.update(survivors)
|
||||
except Exception as ex:
|
||||
LOGGER.exception("fast DB reconciliation failed for %s: %s", r, ex)
|
||||
|
||||
paths: list[str] = []
|
||||
if "models" in roots:
|
||||
paths.extend(collect_models_files())
|
||||
if "input" in roots:
|
||||
@ -113,10 +118,12 @@ async def sync_seed_assets(roots: list[schemas_in.RootType]) -> None:
|
||||
if "output" in roots:
|
||||
paths.extend(list_tree(folder_paths.get_output_directory()))
|
||||
|
||||
processed = 0
|
||||
async with await create_session() as sess:
|
||||
for p in paths:
|
||||
try:
|
||||
if os.path.abspath(p) in existing_paths:
|
||||
skipped_existing += 1
|
||||
continue
|
||||
st = os.stat(p, follow_symlinks=True)
|
||||
if not int(st.st_size or 0):
|
||||
continue
|
||||
@ -134,17 +141,20 @@ async def sync_seed_assets(roots: list[schemas_in.RootType]) -> None:
|
||||
owner_id="",
|
||||
)
|
||||
|
||||
processed += 1
|
||||
if processed % 500 == 0:
|
||||
created += 1
|
||||
if created % 500 == 0:
|
||||
await sess.commit()
|
||||
except OSError:
|
||||
continue
|
||||
await sess.commit()
|
||||
finally:
|
||||
LOGGER.info(
|
||||
"Assets scan(roots=%s) completed in %.3f s",
|
||||
"Assets scan(roots=%s) completed in %.3f s (created=%d, skipped_existing=%d, total_seen=%d)",
|
||||
roots,
|
||||
time.perf_counter() - t_total,
|
||||
created,
|
||||
skipped_existing,
|
||||
len(paths),
|
||||
)
|
||||
|
||||
|
||||
@ -406,7 +416,9 @@ def _append_error(prog: ScanProgress, *, path: str, message: str) -> None:
|
||||
})
|
||||
|
||||
|
||||
async def _fast_db_consistency_pass(root: schemas_in.RootType) -> None:
|
||||
async def _fast_db_consistency_pass(
|
||||
root: schemas_in.RootType, *, collect_existing_paths: bool = False
|
||||
) -> Optional[set[str]]:
|
||||
"""
|
||||
Quick pass over asset_cache_state for `root`:
|
||||
- If file missing and Asset.hash is NULL and the Asset has no other states, delete the Asset and its infos.
|
||||
@ -414,10 +426,12 @@ async def _fast_db_consistency_pass(root: schemas_in.RootType) -> None:
|
||||
* If at least one state for this Asset is fast-ok, delete the missing state.
|
||||
* If none are fast-ok, add 'missing' tag to all AssetInfos for this Asset.
|
||||
- If at least one state becomes fast-ok for a hashed Asset, remove the 'missing' tag.
|
||||
When collect_existing_paths is True, returns a set of absolute file paths
|
||||
that still have a live asset_cache_state row for this root after reconciliation.
|
||||
"""
|
||||
prefixes = prefixes_for_root(root)
|
||||
if not prefixes:
|
||||
return
|
||||
return set() if collect_existing_paths else None
|
||||
|
||||
conds = []
|
||||
for p in prefixes:
|
||||
@ -429,7 +443,7 @@ async def _fast_db_consistency_pass(root: schemas_in.RootType) -> None:
|
||||
|
||||
async with await create_session() as sess:
|
||||
if not conds:
|
||||
return
|
||||
return set() if collect_existing_paths else None
|
||||
|
||||
rows = (
|
||||
await sess.execute(
|
||||
@ -467,6 +481,7 @@ async def _fast_db_consistency_pass(root: schemas_in.RootType) -> None:
|
||||
|
||||
# Apply actions
|
||||
to_set_verify: list[int] = []
|
||||
survivors: set[str] = set()
|
||||
for aid, acc in by_asset.items():
|
||||
a_hash = acc["hash"]
|
||||
states = acc["states"]
|
||||
@ -481,7 +496,10 @@ async def _fast_db_consistency_pass(root: schemas_in.RootType) -> None:
|
||||
asset = await sess.get(Asset, aid)
|
||||
if asset:
|
||||
await sess.delete(asset)
|
||||
# else leave it for the slow scan to verify/rehash
|
||||
else:
|
||||
for s in states:
|
||||
if s["exists"]:
|
||||
survivors.add(os.path.abspath(s["obj"].file_path))
|
||||
else:
|
||||
if any_fast_ok:
|
||||
# Remove 'missing' and delete just the stale state rows
|
||||
@ -490,9 +508,15 @@ async def _fast_db_consistency_pass(root: schemas_in.RootType) -> None:
|
||||
await sess.delete(await sess.get(AssetCacheState, st.id))
|
||||
with contextlib.suppress(Exception):
|
||||
await remove_missing_tag_for_asset_id(sess, asset_id=aid)
|
||||
for s in states:
|
||||
if s["exists"]:
|
||||
survivors.add(os.path.abspath(s["obj"].file_path))
|
||||
else:
|
||||
with contextlib.suppress(Exception):
|
||||
await add_missing_tag_for_asset_id(sess, asset_id=aid, origin="automatic")
|
||||
for s in states:
|
||||
if s["exists"]:
|
||||
survivors.add(os.path.abspath(s["obj"].file_path))
|
||||
for s in states:
|
||||
if s["exists"] and not s["fast_ok"]:
|
||||
to_set_verify.append(s["obj"].id)
|
||||
@ -505,3 +529,4 @@ async def _fast_db_consistency_pass(root: schemas_in.RootType) -> None:
|
||||
)
|
||||
await sess.flush()
|
||||
await sess.commit()
|
||||
return survivors if collect_existing_paths else None
|
||||
|
||||
Loading…
Reference in New Issue
Block a user