DataFlow Documentation
DataFlow helps you unify supplier feeds (XML, CSV, XLSX, JSON APIs) into a stable output endpoint. Use visual mapping, filters, snapshots, and scheduling so your downstream imports stop breaking.
Overview
The platform is built around the idea that messy supplier data should be handled once: connect sources, map fields into a single schema, normalize values, and publish a unified endpoint.
Core concepts
Getting started
- 1) Create a workspaceSign up, then create a workspace for your company/project. Workspaces help isolate data and permissions.
- 2) Add a sourceGo to Sources → New source. Choose URL or file upload, then pick the record list in the extraction wizard (for XML/JSON).
- 3) Create a feedGo to Feeds → New feed. Select the sources that contribute to your unified output and pick a scheduling mode (Live/Hybrid/Cron).
- 4) Map fieldsOpen Mappings. Drag fields from the source preview into target fields. Add normalizations (trim, casing, number parsing) and filters.
- 5) Publish & testOpen the feed page and copy the public endpoint. Use an API key if enabled. Verify that JSON/CSV/XML output is accepted by your downstream system.
Start with a minimal output schema (SKU, title, price, stock). Add fields gradually after you confirm your downstream import works.
Outputs and formats
Each feed exposes a stable public endpoint. Depending on your setup, you can request JSON, CSV, or XML. If API keys are enabled, include the key in an HTTP header.
# JSON
curl -H "X-API-Key: YOUR_KEY" https://YOUR_DOMAIN/f/your-feed-slug.json
# CSV
curl https://YOUR_DOMAIN/f/your-feed-slug.csv
# XML
curl https://YOUR_DOMAIN/f/your-feed-slug.xmlScheduling modes
Filters & normalization
Real supplier feeds contain inconsistent formats: comma decimals, mixed casing, missing brands, or category drift. Use normalization and filters to keep your output consistent.
- • Trim whitespace, normalize casing, remove control characters
- • Parse numbers safely (prices, weights), enforce decimal separator
- • Set defaults (e.g., stock = 0 when missing)
- • Filter out deleted/out-of-stock items at source or unified level
Snapshots & run history
Runs show what happened when the pipeline executed. Snapshots store inputs and outputs so you can compare changes over time.
Webhook exports
Webhook exports let you push the unified output to another system when a run finishes. Typical use cases: updating a cache, triggering downstream imports, or notifying a pipeline.
When a feed run completes, DataFlow sends a JSON body to your configured URL (best-effort — failures do not block the run).
{
"feed": { "id": "...", "slug": "your-feed-slug" },
"generated_at": "2026-02-03T...Z",
"data": [
{ "sku": "...", "name": "...", "price": 123.45 }
]
}Billing, usage & troubleshooting
Billing is enforced by plan limits (feeds, sources, runs/day, features). If you upgrade and don’t see changes: use the Billing page’s Re-sync button (it fetches your Stripe subscription and updates the workspace).
- • Ensure Stripe keys are set in Admin → Billing
- • Ensure you created plan prices (monthly/yearly)
- • If needed, click Re-sync billing on the workspace billing page
- • Missing customer address → open Billing Portal and add address
- • VAT ID collection requires name updates → handled automatically in Checkout
- • Key rotation between Stripe accounts → re-save prices in Admin