Import Media
Queue cloud-sync handover batches; the background worker consumes them.
About this section: Import MediaQueue cloud-sync handover bundles; the worker imports them in the background.
Purpose
Consume the upstream cloud-sync handover bundles from /mnt/s3-sync/assets/sync/ into the global library (asset, asset_gallery, asset_collection) per Decision 639. The CGI never imports inline; it queues batches into import_job for the background worker (Decision 640).
Flow
- Scan: walks the mount for
.readysentinels and records newingest_batchrows. - Pre-flight: validates the manifest shape, sentinel pair, and per-item decisions; no library writes.
- Queue: auto-preflights if needed, then inserts an
import_jobrow. Live progress shows on Import Tasks. - Bulk queue: select multiple batches and queue them in one click; the worker drains the queue in priority + FIFO order.
Example
Scan finds 37e428d2-a4f6-4a31-bacf-0c835fb24084.ready. Tick its checkbox (or several at once), click Queue selected. The worker claims the next queued job within seconds and writes job_* events as it progresses.
Consequences
- The library is site-agnostic. Sites surface library content via
asset_site_membership, gallery/collection membership, content allowlists, and access packages. - Cancel is allowed only while a job is queued; once running, only
systemctl stop gpu-admin-panel-import-workerstops it. - Action
removeonly soft-deletes rows viadeleted_at; nothing is hard-deleted. - Frosted, blurred, and vivid_blurred derivatives are intentionally not persisted.
Deeper dive
Sourced from gpu-floor/MIGRATION-INTRO.md and the queued-worker decision in DECISION_LOG.md.
Worker contract
A long-running import-worker.pl daemon claims queued jobs using SELECT ... FOR UPDATE SKIP LOCKED (MariaDB 10.6+). It updates heartbeat_at every few items; if a worker dies, its running jobs are reset to queued after the configured stale threshold (default 15 minutes).
Sentinel protocol
A bundle is consumable only when both <sync_id>.json and <sync_id>.ready exist and the JSON has status = completed and verification.passed = true. The importer recomputes manifest_sha256 at consumption time so audit history records what we actually ingested.
Source-of-truth mapping
source/source.json->asset_source(cold provenance).source/data/photo.manifest.json+photo.analysis.json+photo.meaning.json+photo.family.json->asset_ai_runrows with explicitanalysis_kind; unknown filenames are captured withsource_data:<file>.- Bundled
cover_hash->asset_coverlinked to the parent asset. - Containers of type
gallery->asset_gallery+asset_gallery_itemjunctions.
Idempotency + integrity
Every multi-row write is wrapped in AdminDB::transaction. Hash-based UNIQUE keys make re-running the same batch a no-op. Per-item failures isolate to ingest_batch_item.decision = failed with last_error; the batch becomes partial when any siblings still succeeded.
Read-only mount + ledger boundary
The /mnt/s3-sync mount is read-only by configuration. The importer never writes to upstream or deletes .ready sentinels; all consumer state lives in ingest_batch / ingest_batch_item / ingest_event / import_job in admin_panel.
Batch a09117d8-8013-4a85-8c57-b2ccdb3d36f6
Status: imported; action: update
Type
event
Action
update
Status
imported
Items
2
Bytes
0
Source host
flux-1
Manifest sha256
74e24cc0da6fa38c...
Started
2026-04-30 00:34:03
Completed
2026-04-30 00:34:04
Manifest summary
- created_at: 2026-04-12T21:35:35Z
- completed_at: 2026-04-12T21:35:35Z
- verification.passed: true
- verification.total_files: 0
- prior_sync_id: bc02ed61-9d39-4538-952a-ed4ced680e96
Job history 1
Newest first. The active job (if any) is at the top.
| id | state | requested_by | queued | claimed | worker | progress | finished | last_error |
|---|---|---|---|---|---|---|---|---|
| 6 | done | - | 2026-04-30 00:34:03 | 2026-04-30 00:34:03 | vultr/pid:356322 | 2/2 | 2026-04-30 00:34:04 | - |
Items 2
Per-item decisions captured by preflight and updated by import. Re-running pre-flight is safe.
| asset_type | asset_hash | cover_hash | decision | target ids | last_error | updated_at |
|---|---|---|---|---|---|---|
| video | 212a6c7a8629cd00... | - | updated | asset:66 | - | 2026-04-30 00:34:04 |
| cover | 707a5b21ccd44b54... | - | linked | asset:66 | - | 2026-04-30 00:34:04 |
Recent events
Append-only audit ledger for this batch (latest 50).
| event_type | payload | created_at |
|---|---|---|
| job_finished | {"error":null,"job_id":"6","phase":"job_finish","summary":{"items_seen":2,"ok":1,"status":"imported","summary":{"ok":1,"remaining":0,"status":"imported","tally":{"created":0,"deferred":0,"failed":0,"l... | 2026-04-30 00:34:04 |
| import_finished | {"phase":"import","remaining":0,"skipped_terminal":0,"status":"imported","tally":{"created":0,"deferred":0,"failed":0,"linked":1,"skipped":0,"updated":1}} | 2026-04-30 00:34:04 |
| item_linked | {"cover_hash":null,"error":null,"hash":"707a5b21ccd44b548145b286df183baa","phase":"item_import","target_ids":{"asset_id":"66","cover_hash":"707a5b21ccd44b548145b286df183baa","parent_hash":"212a6c7a862... | 2026-04-30 00:34:04 |
| item_updated | {"cover_hash":null,"error":null,"hash":"212a6c7a8629cd00baae3940aa717e6c","phase":"item_import","target_ids":{"asset_id":"66"},"type":"video"} | 2026-04-30 00:34:04 |
| import_started | {"action":"update","already_terminal_count":0,"container_count":1,"item_count":2,"manifest_sha256":"74e24cc0da6fa38c3985a67239c823a05d152e0aac76122702b86b0354ded989","max_items":0,"phase":"import"} | 2026-04-30 00:34:03 |
| preflight_finished | {"counts":{"will_create":0,"will_fail":0,"will_link":1,"will_skip":0,"will_update":1},"failure_samples":[],"item_count":2,"item_type_counts":{"cover":1,"video":1},"ok":1,"phase":"preflight"} | 2026-04-30 00:34:03 |
| preflight_started | {"action":"update","container_count":1,"item_count":2,"item_type_counts":{"cover":1,"video":1},"manifest_sha256":"74e24cc0da6fa38c3985a67239c823a05d152e0aac76122702b86b0354ded989","phase":"preflight",... | 2026-04-30 00:34:03 |
| job_claimed | {"job_id":"6","phase":"job_claim","priority":100,"worker_id":"vultr/pid:356322"} | 2026-04-30 00:34:03 |
| job_reset | {"job_id":"6","phase":"job_reset","prior_claimed_by":"simulated-crashed-worker/pid:99999","prior_heartbeat_at":"2026-04-30 00:04:03","reaped_by":"vultr/pid:356322","threshold_seconds":"900"} | 2026-04-30 00:34:03 |
| scan | {"source": "filesystem", "manifest_sha256": "74e24cc0da6fa38c3985a67239c823a05d152e0aac76122702b86b0354ded989", "sync_type": "event", "action": "update", "item_count": "2", "total_bytes": "0", "manife... | 2026-04-30 00:29:20 |