🧭 Quick Return to Map
You are in a sub-page of Automation Platforms.
To reorient, go back here:
- Automation Platforms — stabilize no-code workflows and integrations
- WFGY Global Fix Map — main Emergency Room, 300+ structured fixes
- WFGY Problem Map 1.0 — 16 reproducible failure modes
Think of this page as a desk within a ward.
If you need the full triage and all prescriptions, return to the Emergency Room lobby.
A focused repair guide for teams building pipelines with Parabola. Goal is simple: stop silent data drift, schema breaks, pagination traps, and idempotency bugs without changing your infra. Use the steps and acceptance targets below to make the fix repeatable.
- A quick path to locate the failing layer in your Parabola flow: input → transform → join → export → webhook/API.
- Structural fixes that survive retries, partial failures, and schema changes.
- Exact links into the WFGY Problem Map where the permanent patch lives.
- CSVs import but downstream counts are off.
- A join explodes row counts or drops keys.
- Pagination or rate limits make exports flaky.
- Webhook tasks replay and create duplicates.
- Column names change and flows keep “succeeding”.
- Schedules “succeed” yet the destination is stale.
-
Data contracts and citations for rows and fields: Data Contracts
-
Live monitoring and run-debug checklists for pipelines: Live Monitoring for RAG and Pipelines · Debug Playbook
-
Boot order and “first run” failures that look like Parabola bugs: Bootstrap Ordering · Deployment Deadlock · Pre-Deploy Collapse
-
Vectorstore and retrieval acceptance targets you may export into: Retrieval Playbook · Retrieval Traceability
-
Lock a data contract for every flow edge
- Define required columns, types, nullability, and primary key.
- Put the contract in the flow description and in a sidecar
.json. - Reject on contract break, do not “coerce”.
-
Make writes idempotent
- Add an idempotency key from source primary key + run id.
- Upsert on key. Soft-delete on tombstone streams.
-
Tame pagination and rate limits
- Use explicit page cursors where available.
- Backoff with jitter and a cap. Persist last good cursor.
- Fail closed on partial pages, resume from cursor.
-
Stabilize joins
- Pre-dedupe on join keys.
- Count rows before and after. Warn if ratio not in [0.9, 1.1] unless configured.
- For one-to-many, aggregate first, then join.
-
Quarantine bad rows
- Sink violations to a “dead-letter” sheet with reason code.
- Never drop silently.
-
Schedule with proof
- Record run hash = inputs’ checksums + step graph rev.
- A run is “good” only if the same hash reproduces.
| Symptom | Root cause | Open this fix |
|---|---|---|
| Row counts drift after CSV import | Type coercion and null handling change silently | Data Contracts |
| Duplicates after webhook retries | No idempotency key on destination | Debug Playbook |
| Join multiplies rows unexpectedly | Non-unique keys or many-to-many join | Live Monitoring |
| Exports fail intermittently | Pagination or rate-limit handling missing | Debug Playbook |
| First run looks “green” but index is empty | Boot order wrong, destination not ready | Bootstrap Ordering |
| Scheduled run “succeeds” but target stale | No acceptance gates or version checks | Live Monitoring |
| Downstream retrieval pulls wrong docs | Snippet schema absent, traceability missing | Retrieval Traceability |
- Inputs: file counts and checksums logged.
- Contract: columns, types, PK declared and enforced.
- Dedupe: before join, after import.
- Idempotency: deterministic key on write path.
- Pagination: cursor persisted between attempts.
- Quarantine: every rejection is stored with reason.
- Acceptance: target store has post-write assertions.
I uploaded TXT OS and Problem Map.
Context: Parabola pipeline failing.
- symptom: [brief]
- sources: [csv/api names]
- current guards: [contract? idempotency? pagination? join?]
Tell me:
1) which layer is failing and why,
2) which exact WFGY page to open,
3) the smallest patch to make writes idempotent and schema-locked,
4) how to verify with row counts and hashes.
Use BBMC/BBPF/BBCR/BBAM if relevant.- Contract violations are zero.
- Duplicate writes are zero across retries.
- Join ratio stays within configured band or the run blocks.
- Pagination resumes from last cursor with no missing pages.
- Destination post-write assertions pass on every schedule.
- Re-running with same inputs reproduces identical hash.
| Tool | Link | 3-Step Setup |
|---|---|---|
| WFGY 1.0 PDF | Engine Paper | 1️⃣ Download · 2️⃣ Upload to your LLM · 3️⃣ Ask “Answer using WFGY + <your question>” |
| TXT OS (plain-text OS) | TXTOS.txt | 1️⃣ Download · 2️⃣ Paste into any LLM chat · 3️⃣ Type “hello world” — OS boots instantly |
| Layer | Page | What it’s for |
|---|---|---|
| ⭐ Proof | WFGY Recognition Map | External citations, integrations, and ecosystem proof |
| ⚙️ Engine | WFGY 1.0 | Original PDF tension engine and early logic sketch (legacy reference) |
| ⚙️ Engine | WFGY 2.0 | Production tension kernel for RAG and agent systems |
| ⚙️ Engine | WFGY 3.0 | TXT based Singularity tension engine (131 S class set) |
| 🗺️ Map | Problem Map 1.0 | Flagship 16 problem RAG failure taxonomy and fix map |
| 🗺️ Map | Problem Map 2.0 | Global Debug Card for RAG and agent pipeline diagnosis |
| 🗺️ Map | Problem Map 3.0 | Global AI troubleshooting atlas and failure pattern map |
| 🧰 App | TXT OS | .txt semantic OS with fast bootstrap |
| 🧰 App | Blah Blah Blah | Abstract and paradox Q&A built on TXT OS |
| 🧰 App | Blur Blur Blur | Text to image generation with semantic control |
| 🏡 Onboarding | Starter Village | Guided entry point for new users |
If this repository helped, starring it improves discovery so more builders can find the docs and tools.