Orchestration layer around the jrunner Java JDBC CLI, replacing the previous shell-based sync system in .archive/pre-rewrite. Includes the FastAPI + Jinja web frontend, per-driver adapters (DB2, MSSQL, PG), wizard-driven module creation with editable dest types and source-sourced table/column descriptions, watermark/hook CRUD, and the engine that runs modules end-to-end. Co-Authored-By: Claude Opus 4.7 <noreply@anthropic.com>
26 KiB
Pipekit — Spec
This spec was built from a clean-slate conversation that rederived the
design from first principles. The previous version is archived at
SPEC_v1_archive.md for reference.
Status
Spec is done. Ready to move to implementation planning.
One item is intentionally deferred: the migration plan for bringing
over the ~90 existing modules from /opt/sync. Not needed to start
implementation — Pipekit can be built and tested against new modules
first, and migration can happen later (likely via a parser that walks
/opt/sync/*/, extracts pull.sql / insert.sql / shell wrapper,
infers merge strategy and key, and creates module rows).
How we got here
Started by asking what was painful about the existing shell-script-based
sync setup. Three things surfaced: authoring new modules is tedious,
observability is poor (no easy way to see what ran, how long, how many
rows, any errors), and there's no central management UI. That framed
Pipekit as an orchestration layer on top of the existing jrunner JDBC
tool — not replacing jrunner, wrapping it with the state and UI that
shell scripts can't provide.
Everything in this document was worked out by walking through concrete
examples from the current /opt/sync modules (code, qcrh,
ffsbglr1) and asking "what would this look like under the new system?"
When the original spec proposed something that didn't fit (like
"watermark is a single column name"), we redesigned it. The result is a
spec that reflects the actual complexity of real modules, not an
idealized simple-sync model.
Motivation
User has ~90 sync modules in /opt/sync today, organized as shell scripts
that wrap jrunner (a JDBC bulk-transfer CLI at /opt/jrunner). Pain points
that drove this redesign:
- Authoring is tedious. Building SQL for new sync modules takes too long — hand-writing pull.sql, insert.sql, the .sh wrapper, the import table DDL.
- No observability. Hard to answer: how often does each module run, how many rows transfer, what SQL was used, when's the next run, how long does it take, are tables in a good state, were there errors on the last run and for which modules.
- No central management. Want a TUI like lazygit for browsing, inspecting, running, configuring modules. User browses with nvim today and wants the TUI to feel as spatial and navigable as a file tree.
What jrunner does (and doesn't)
jrunner (at /opt/jrunner) is a Java CLI that does two things:
- Migration mode — given source connection (
-scu/-scn/-scp), dest connection (-dcu/-dcn/-dcp), a SQL file (-sq), and a dest table (-dt), it streams rows from source to dest with batched INSERTs. - Query mode — same source flags but no dest flags, outputs query results to stdout in CSV/TSV. Useful for piping to visidata, less, etc.
It has no merge logic, no scheduling, no state, no awareness of incremental syncs. It's a dumb pipe. That's the right shape — Pipekit wraps it with the orchestration layer.
Architecture
jrunner (Java CLI — bulk JDBC transfer + query mode)
↑
engine (Python — orchestrates jrunner, watermarks, merge, hooks, run log)
↑
API (FastAPI — REST, Basic Auth)
↑
TUI / web UI / curl
The engine shells out to jrunner for everything that touches a database — bulk transfers, watermark resolver queries, hooks. No separate JDBC layer in Python. One driver-loading code path, one set of bugs.
The API exists so a web front-end or curl can drive Pipekit, not just the TUI.
Storage: SQLite
Everything lives in one SQLite file (pipekit.db). Why:
- ~90+ modules already exists; flat files don't scale to "show me all modules that errored last night" type queries.
- The SQL itself belongs in the database, not as file references — a module is a self-contained unit and splitting it across rows + files means two things to keep in sync.
- Single file, copy with
cp, no server. Schema translates to PostgreSQL later if ever needed.
User was uneasy about losing filesystem browsing. Resolution: the TUI is the
file browser. Inspecting a module feels like cat, editing opens $EDITOR,
the module list feels like ls. For raw access, sqlite3 pipekit.db works.
Module model
A module = one sync job. Fields:
namesource_connection_id,dest_connection_iddest_tablestaging_table(auto-managed:pipekit_staging.{name})source_query— full SQL text with{watermark_name}placeholders. Free-form.merge_strategy—full/incremental/appendmerge_key— destination column(s) used in DELETE before INSERTenabledrunning(lock flag — see locking section)
The source query is a text blob. Not split into structured columns. The
TUI offers a column-editor mode that parses the SELECT list out of the
stored query, lets you edit it as a table, and splices the new SELECT list
back in (preserving CTEs, FROM, WHERE). For queries the parser can't handle
(too complex), the TUI falls back to raw $EDITOR. Raw editing always
works.
Merge strategies
Two patterns from existing scripts:
- full — TRUNCATE dest, INSERT all from staging
- incremental — pull delta via watermark, DELETE rows in dest matching merge_key, INSERT from staging
- append — INSERT only, no deletes
No upsert. The DELETE+INSERT approach already handles row-level changes without needing column-by-column ON CONFLICT UPDATE SET clauses.
Watermarks (multi, type-agnostic, resolver SQL)
A module can have multiple named watermarks. Real example from user: a
query that needs both {date} (max modified-timestamp from one table) and
{number} (max order number from another) to build a list of changed orders
to repull.
A watermark =
name— placeholder name in the source queryconnection_id— which connection runs the resolver (could be dest, source, or a third)resolver_sql— free-form SQL. Engine runs it via jrunner query mode, takes first row's first column as a string.default_value— used if resolver returns NULL or zero rows
Type-agnostic. The engine reads the resolver result as an opaque string and
substitutes it literally. No type coercion. The user controls quoting in the
resolver SQL itself (e.g. wrap in quote_literal() if you want '2610',
return raw if you want 2610).
Dialect-aware by user. The user writes the resolver in the connection's dialect. Engine doesn't translate. Same as today — they already write DB2 in pull.sql and PG in insert.sql.
No hidden generation. Resolved SQL gets materialized before each run
and stored on the module record (next_resolved_query or similar) so the TUI
can always show "here's exactly what would run next." After the run, the
exact resolved SQL goes into the run_log.
Hooks
A module can have post-execution hooks for things like
REFRESH MATERIALIZED VIEW rlarp.cust or CALL rlarp.osm_stack_refresh().
A hook =
module_id,run_orderconnection_id— usually dest, but anywheresqlrun_on—success/failure/always
Hooks run sequentially after the merge. Failures get logged but don't roll back the merge (it's already committed).
No group-level hooks for now. Decision deferred. The REFRESH MATERIALIZED VIEW rlarp.cust at the end of codes.sh would attach to whichever module
logically owns that data, even if not strictly the last in order. Add group
hooks later if it gets painful.
Engine flow (per module run)
- Acquire lock atomically:
UPDATE module SET running=1 WHERE id=? AND running=0. If row count is 0, bail with "already running." - Resolve watermarks. For each watermark: shell out to jrunner query mode against the watermark's connection with its resolver SQL. Take first row's first column as a string. Fall back to
default_valueon NULL/empty. - Materialize the resolved source query. Substitute
{name}placeholders insource_query. Store on the module record so the TUI can preview. - Truncate staging (
TRUNCATE pipekit_staging.{module_name}). - Run jrunner (migration mode) with the resolved query, target = staging.
- Materialize the merge SQL based on strategy + merge_key.
- Run merge against dest connection (also via jrunner, or whatever path the engine uses for SQL execution).
- Run hooks in order, respecting
run_on. - Write
run_logentry with everything (see below). - Release lock in a
finallyblock — always runs, even on error.
Locking
The running flag on the module is the lock. The atomic UPDATE-with-WHERE
above ensures no race window. Belt-and-suspenders for stuck locks:
- PID-based. Store the API process PID/UUID on the lock. On API startup, clear locks owned by PIDs that no longer exist.
- Time-based backstop. On startup, also clear locks held longer than some absurd threshold (e.g. 24h).
Lock is enforced regardless of trigger source — scheduler, group runner, ad-hoc single-module, ad-hoc group run. All paths hit the same atomic check.
No separate group lock needed. If a group runner tries to start a module that's already locked, it fails on that module and stops the group (per stop-on-failure rule).
Run log / observability
Two tables:
group_run(
id, group_id, started_at, finished_at, status, triggered_by
-- triggered_by: schedule | manual | null
)
run_log(
id,
module_id,
group_run_id, -- nullable; set when run as part of a group
started_at, finished_at,
row_count,
status, -- running | success | error | cancelled
error,
resolved_source_sql, -- exact SQL that ran on source
merge_sql, -- exact merge SQL that ran on dest
watermark_values_json, -- {prev_period: "'2610'", ...}
jrunner_stdout,
jrunner_stderr,
hook_log
)
Module history is independent of group context — WHERE module_id=? shows
every run, scheduled or manual, group or standalone. The group_run_id is
just an annotation.
Run detail screen (in TUI) shows: timing, status, row count, trigger
context, watermark values, plus keys to open in $EDITOR:
s— resolved source SQLm— merge SQLh— hook outputo— jrunner stdout/stderr
Global run log (L from main screen) — sortable, filterable across all
modules and groups. Answer "show me everything that errored in the last 24
hours" in two keystrokes.
Groups and scheduling
grp(id, name)
group_member(id, group_id, module_id, run_order)
-- many-to-many; same module can live in multiple groups with different run_orders
schedule(id, group_id, cron_expr, enabled)
-- a group can have 0..N schedules
Sequential execution, stop on failure. Mirrors the set -e behavior of
existing orchestrator scripts.
Many-to-many membership. Junction table is needed anyway for run_order,
so many-to-many costs nothing extra. Unique constraint can be added later if
ever needed.
Schedule attaches to groups, not modules. Matches the user's mental model and avoids a huge cron-list. Individual modules can still be run ad-hoc.
Scheduler. Background thread inside the API process. Wakes every minute,
evaluates all enabled schedules, fires any whose cron matches. A scheduled
fire and a manual fire use the same code path — only triggered_by differs.
Ad-hoc runs:
POST /modules/{id}/run— single modulePOST /groups/{id}/run— whole group sequentially
Both create normal run_log entries.
Connections and credentials
driver(id, name, jar_file, class_name, url_template)
connection(
id,
name,
driver_id,
jdbc_url,
username,
password,
default_dest_connection_id, -- nullable; wizard default when this is source
default_dest_schema, -- nullable; wizard default when this is source
notes,
created_at, updated_at
)
Credentials = env var references. The password column stores something
like $DB2PW. Engine resolves at runtime by reading the env var. Passwords
never live in the database. Matches existing setup
(/opt/sync/.env + shell scripts) and keeps pipekit.db safe to copy/back-up.
Test-connection: engine runs a trivial query (SELECT 1 or equivalent) via
jrunner against the connection. Confirms URL, credentials, driver all work.
jrunner handles all SQL execution — bulk transfers (migration mode) and single-value queries for watermark resolvers / hooks (query mode). Trade-off: ~100ms JVM spawn per resolver call, but one tool, one set of bugs, one driver-loading path.
Bootstrap / install hygiene
Pipekit verifies jrunner exists on startup (configurable path in
config.yaml). If missing, surfaces a clear error pointing at
/opt/jrunner/deploy.sh.
pipekit doctor CLI command — checks jrunner present, jrunner version,
drivers loadable, database accessible, all configured connections testable.
First thing to run after a git pull.
Packaging. Start loose-coupled (install jrunner separately, point Pipekit at it). Bundle later if/when the two-step gets annoying.
New module wizard
The centerpiece for fixing the authoring pain. Goal: from "I want to sync table X from connection Y" to "module created, query previewed, ready to test-run" in under a minute.
Step 1 — Source
Pick source connection. Filter by schema. Search tables incrementally. The
TUI calls jrunner in query mode against the source's INFORMATION_SCHEMA
equivalent (DB2: SYSIBM.SYSTABLES, SQL Server / PG: INFORMATION_SCHEMA.TABLES).
Step 2 — Columns
The engine introspects the chosen table. Proposes one row per column with:
- In/out toggle (default all on; toggle off the noise like
dcfut*futures) - Default alias — lowercase, special chars stripped:
DCORD#→dcord - Default source expression — bare column for most types;
RTRIM(col)for char/varchar;CASE WHEN col IN ('0001-01-01','9999-12-31') THEN NULL ELSE col ENDfor date (sentinel-NULL pattern from existing modules) - Default dest type — mapped from source:
INT→integer,DECIMAL(15,4)→numeric(15,4),CHAR(40)→text,DATE→date, etc.
e opens an edit modal for one row to override alias / expression / type.
Most of the time you accept defaults.
Step 3 — Destination & merge
Pick dest connection. Dest table defaults to
{source_conn.default_dest_schema}.{lowercase_source_table_name}. Pick
merge strategy. Pick merge key from a dropdown of dest column names. Add
zero or more watermarks via a sub-form.
Multiple destinations are real (e.g. PG → SQL Server). The wizard
doesn't assume one dest. Each source connection has a
default_dest_connection_id + default_dest_schema pair that
pre-populate Step 3. Both are nullable; fallback is last-used dest.
Step 1 — Source (driver-dependent browse form)
Different drivers need different scope fields ("qualifiers") to identify a
table. DB2 needs just schema. SQL Server can need up to three:
linked_server, database, schema (any combination — linked server
optional, database optional, schema defaults to dbo). This is because
SQL Server can reference tables in other databases on the same server, or
tables on entirely different servers via linked servers — and the FROM
clause syntax changes (schema.table, db.schema.table,
[linked].[db].[schema].[table]).
Each driver exposes:
class Driver:
def browse_fields(self) -> list[BrowseField]:
"""Qualifier fields for the wizard's Step 1 form."""
def list_tables(self, **qualifiers) -> list[Table]:
"""INFORMATION_SCHEMA query using whatever qualifiers are set."""
def get_columns(self, table_name: str, **qualifiers) -> list[Column]:
"""Column lookup for a specific table."""
def qualified_table_name(self, table_name: str, **qualifiers) -> str:
"""FROM-clause identifier. Wizard-time only."""
def map_type(self, source_type) -> str: ...
def default_expression(self, source_type, column_name) -> str: ...
def quote_identifier(self, name) -> str: ...
Textual renders Step 1 dynamically from browse_fields(). The wizard
calls qualified_table_name() once to bake the FROM clause into the
stored source query. Linked servers / qualifiers are not first-class in
Pipekit — they exist only as syntax inside the generated FROM. Nothing
is persisted on the module about how the table was qualified at author
time. If you later need to add a column, you type the expression and
alias by hand in the column editor — no re-browsing needed.
Step 4 — Preview
Show the generated source query, generated staging DDL, generated merge SQL.
Everything visible. e to drop into $EDITOR for free-form fixes. c to
create — writes the module row, creates the staging table on dest, offers a
test-run.
Per-driver capability needed
Each driver module (engine/drivers/db2.py, etc.) implements:
list_tables(schema_filter)— SQL template for INFORMATION_SCHEMAget_columns(schema, table)— column name, type, length, nullablemap_type(source_type)→ dest typedefault_expression(source_type, column_name)→ wrap in RTRIM, CASE, etc.quote_identifier(name)—"DCORD#"(DB2/PG) vs[DCORD#](MSSQL)
Defaults are opinions hardcoded in driver modules for now. Lift to a
driver_default table later if configurability is ever needed.
Wizard scope (what it does NOT do)
- No CTE-based queries. Wizard generates simple
SELECT cols FROM table WHERE watermark. For complex queries (likeffsbglr1), create with the wizard and edit the source query post-creation viae. - No multi-watermark wizard. Single watermark. Add more after.
- No hooks in the wizard. Add hooks from the module detail screen.
- No group assignment in the wizard. Assign separately.
These are intentional. The wizard handles the 80% case fast. The 20% cases are post-creation edits where you already have a working module to start from.
TUI — main screen sketch
Pipekit
─────────────────────────────────────────────────
▼ s7830956 (AS/400 DB2)
✔ code full 2m ago 1,204 rows 0.8s
✔ name full 2m ago 892 rows 0.6s
✔ qcrh incr 2m ago 1,031 rows 3.2s
✗ qcri incr 2m ago — err
○ cust full disabled
▼ usmidsql01 (SQL Server)
✔ live_quotes full 2m ago 340 rows 1.1s
Groups
pricing 9 modules cron 0 20 2 * * * next: 2:20am
codes 26 modules cron 0 0 2 * * * next: 2:00am
Modules grouped by source connection (mirrors today's directory layout). Status / strategy / last-run / row-count / duration on each line. Groups at the bottom with schedules and next-fire times.
i inspect, r run, l history, L global log, n new module, c
connections, / search, j/k navigate, q quit. Should feel like lazygit /
nvim file tree.
Module detail (i)
Top: module info (strategy, merge key, watermark, dest table, staging table, enabled, last/next run). Middle: column table (parsed from source query). Bottom: keybindings.
Keys open things in $EDITOR (read-only):
q— next resolved source SQLm— merge SQLb— base query template (with placeholders)e— edit base query (writable)w— watermarksh— hooksc— column editor (parsed from query)r— runl— history
API surface
REST over HTTP, FastAPI, HTTP Basic Auth on all endpoints except
/health. In practice the API only uses GET (reads) and POST
(writes) — PUT/DELETE avoided to keep the mental model simple.
Resource CRUD
Every core table (connection, driver, module, watermark, hook, group, group_member, schedule) gets the same URL pattern:
GET /things list (with filter query params)
GET /things/{id} read one
POST /things create
POST /things/{id} update
POST /things/{id}/delete delete
JSON shape = snake_case matching database columns. ISO 8601 timestamps. Integer IDs. No transformation layer between SQL and JSON.
Operation endpoints
Anything with side effects or that composes multiple steps:
POST /connections/{id}/test run SELECT 1 via jrunner; return ok/fail/elapsed
GET /modules/{id}/preview return next resolved source SQL + merge SQL
(runs watermark resolvers but does NOT sync)
GET /modules/{id}/columns parse source query, return column list
POST /modules/{id}/run start async run; return {run_id} immediately
POST /groups/{id}/run start async group run; return {group_run_id}
POST /modules/{id}/cancel cancel running module (release lock, kill jrunner)
POST /groups/{id}/cancel cancel running group
GET /runs list runs (filter: ?module_id= ?status= ?since=)
GET /runs/{id} run detail (SQL, stdout/stderr, hook output)
GET /runs/{id}/stream Server-Sent Events: live log + status
GET /group-runs list group runs
GET /group-runs/{id} group run with child module runs
GET /modules/{id}/runs shortcut: runs for one module
Introspection endpoints (wizard backend)
POST /introspect/tables body: {connection_id, qualifiers: {...}}
POST /introspect/columns body: {connection_id, table_name, qualifiers}
POST /introspect/propose body: {connection_id, table_name, qualifiers}
returns a ready-to-POST module JSON
propose is curl-able — you can generate a module proposal, tweak the
JSON, then POST it to /modules to create. No TUI required.
System endpoints
GET /health liveness only, no auth required
GET /doctor full check (jrunner, drivers, db, connections, scheduler)
powers `pipekit doctor` CLI
GET /settings
POST /settings/{key}
Async runs + SSE
POST /modules/{id}/run does NOT block. It atomically acquires the
module lock, kicks off the sync in a background task, and returns
{"run_id": 4892} immediately.
Two ways to watch a run after that:
- Polling —
GET /runs/{id}returns the run_log row; keep hitting it untilstatus != running. Simple, works anywhere. - Streaming —
GET /runs/{id}/streamopens a Server-Sent Events connection. The server pushes event lines as things happen — log lines, row-count updates, final status. The TUI uses this for the run watch screen. curl supports it with-N(no buffering).
SSE is plain HTTP with a long-lived connection, not WebSockets. Simpler
to implement, works in browsers natively (EventSource in JS), works in
curl for debugging.
Splitting start from watch (two endpoints) means:
- Cron-triggered runs don't have to watch
- Curl scripting can fire-and-forget
- TUI can reconnect to an already-running sync if it crashes mid-run
Auth
HTTP Basic. Username/password in the settings table. Single-user tool
for now; swap to JWT later if multi-user is ever needed, without
breaking URL structure.
TUI = HTTP client
The TUI never touches SQLite directly. Every screen reads from an endpoint. This guarantees zero behavioral drift between TUI and any future web UI, and makes the API the single source of truth for behavior.
Open questions still to answer
Wizard defaults match user's mental model?Confirmed — RTRIM, sentinel-date NULL, lowercased aliases are fine for now.Dest table default?Resolved — per-source connectiondefault_dest_connection_id+default_dest_schema.API surface.Resolved — REST, GET/POST only, async runs, SSE for live output, CRUD + operations + introspection mix.- Migration plan. Deferred. Would involve a parser that walks
/opt/sync/*/, extracts pull.sql / insert.sql / sh wrapper, infers merge strategy and key, creates module rows.
Decisions log (fast reference)
| Decision | Choice |
|---|---|
| Storage | SQLite, single file |
| Where SQL lives | In the database (text blobs), not files |
| Source query shape | Free text with {watermark} placeholders |
| Columns | Parsed from query; not separate rows; wizard auto-introspects on create |
| Watermarks | Multiple per module, type-agnostic, free-form resolver SQL |
| Merge strategies | full / incremental / append (no upsert) |
| Hooks | Per-module, post-merge, run_on success/failure/always |
| Group hooks | Deferred — not needed yet |
| Group membership | Many-to-many (junction table for run_order anyway) |
| Group execution | Sequential, stop on failure |
| Schedules | Attach to groups; multiple schedules per group allowed |
| Locking | Atomic UPDATE on module.running; PID + time-based stale clearing |
| Credentials | Env var references ($DB2PW); resolved at runtime |
| SQL execution | Everything via jrunner (migration + query mode) |
| Materialized SQL | Always — resolved source SQL stored before run + after run |
| Install | Loose-coupled to jrunner for now; bundle later |
| TUI feel | Like lazygit / nvim file tree; spatial, keyboard-driven |
| Authoring | Wizard handles 80% case; post-creation editing handles the rest |
| Multiple destinations | Supported. Source conn holds default_dest_connection_id + default_dest_schema for wizard prepopulation |
| Driver browse fields | Per-driver qualifier set (schema for DB2/PG, up to linked_server/database/schema for MSSQL) |
| Linked servers | Not first-class; only affect FROM-clause syntax at author time; not persisted on module |
| API style | REST, GET for reads, POST for writes, no PUT/DELETE |
| Run model | Async — POST /run returns run_id immediately; watch via polling or SSE stream |
| Live output | Server-Sent Events (SSE) — plain HTTP, curl-friendly, browser-native |
| Auth | HTTP Basic, single user, creds in settings table |
| TUI ↔ backend | TUI is an HTTP client; never touches SQLite directly |