mirror of
https://github.com/koala73/worldmonitor.git
synced 2026-04-25 17:14:57 +02:00
* feat(seed-contract): PR 2a — runSeed envelope dual-write + 91 seeders migrated
Opt-in contract path in runSeed: when opts.declareRecords is provided, write
{_seed, data} envelope to the canonical key alongside legacy seed-meta:*
(dual-write). State machine: OK / OK_ZERO / RETRY with zeroIsValid opt.
declareRecords throws or returns non-integer → hard fail (contract violation).
extraKeys[*] support per-key declareRecords; each extra key writes its own
envelope. Legacy seeders (no declareRecords) entirely unchanged.
Migrated all 91 scripts/seed-*.mjs to contract mode. Each exports
declareRecords returning the canonical record count, and passes
schemaVersion: 1 + maxStaleMin (matched to api/health.js SEED_META, or 2.5x
interval where no registry entry exists). Contract conformance reports 84/86
seeders with full descriptor (2 pre-existing warnings).
Legacy seed-meta keys still written so unmigrated readers keep working;
follow-up slices flip health.js + readers to envelope-first.
Tests: 61/61 PR 1 tests still pass.
Next slices for PR 2:
- api/health.js registry collapse + 15 seed-bundle-*.mjs canonicalKey wiring
- reader migration (mcp, resilience, aviation, displacement, regional-snapshot)
- direct writers — ais-relay.cjs, consumer-prices-core publish.ts
- public-boundary stripSeedEnvelope + test migration
Plan: docs/plans/2026-04-14-002-fix-runseed-zero-record-lockout-plan.md
* fix(seed-contract): unwrap envelopes in internal cross-seed readers
After PR 2a enveloped 91 canonical keys as {_seed, data}, every script-side
reader that returned the raw parsed JSON started silently handing callers the
envelope instead of the bare payload. WoW baselines (bigmac, grocery-basket,
fear-greed) saw undefined .countries / .composite; seed-climate-anomalies saw
undefined .normals from climate:zone-normals:v1; seed-thermal-escalation saw
undefined .fireDetections from wildfire:fires:v1; seed-forecasts' ~40-key
pipeline batch returned envelopes for every input.
Fix: route every script-side reader through unwrapEnvelope(...).data. Legacy
bare-shape values pass through unchanged (unwrapEnvelope returns
{_seed: null, data: raw} for any non-envelope shape).
Changed:
- scripts/_seed-utils.mjs: import unwrapEnvelope; redisGet, readSeedSnapshot,
verifySeedKey all unwrap. Exported new readCanonicalValue() helper for
cross-seed consumers.
- 18 seed-*.mjs scripts with local redisGet-style helpers or inline fetch
patched to unwrap via the envelope source module (subagent sweep).
- scripts/seed-forecasts.mjs pipeline batch: parse() unwraps each result.
- scripts/seed-energy-spine.mjs redisMget: unwraps each result.
Tests:
- tests/seed-utils-envelope-reads.test.mjs: 7 new cases covering envelope
+ legacy + null paths for readSeedSnapshot and verifySeedKey.
- Full seed suite: 67/67 pass (was 61, +6 new).
Addresses both of user's P1 findings on PR #3097.
* feat(seed-contract): envelope-aware reads in server + api helpers
Every RPC and public-boundary reader now automatically strips _seed from
contract-mode canonical keys. Legacy bare-shape values pass through unchanged
(unwrapEnvelope no-ops on non-envelope shapes).
Changed helpers (one-place fix — unblocks ~60 call sites):
- server/_shared/redis.ts: getRawJson, getCachedJson, getCachedJsonBatch
unwrap by default. cachedFetchJson inherits via getCachedJson.
- api/_upstash-json.js: readJsonFromUpstash unwraps (covers api/mcp.ts
tool responses + all its canonical-key reads).
- api/bootstrap.js: getCachedJsonBatch unwraps (public-boundary —
clients never see envelope metadata).
Left intentionally unchanged:
- api/health.js / api/seed-health.js: read only seed-meta:* keys which
remain bare-shape during dual-write. unwrapEnvelope already imported at
the meta-read boundary (PR 1) as a defensive no-op.
Tests: 67/67 seed tests pass. typecheck + typecheck:api clean.
This is the blast-radius fix the PR #3097 review called out — external
readers that would otherwise see {_seed, data} after the writer side
migrated.
* fix(test): strip export keyword in vm.runInContext'd seed source
cross-source-signals-regulatory.test.mjs loads scripts/seed-cross-source-signals.mjs
via vm.runInContext, which cannot parse ESM `export` syntax. PR 2a added
`export function declareRecords` to every seeder, which broke this test's
static-analysis approach.
Fix: strip the `export` keyword from the declareRecords line in the
preprocessed source string so the function body still evaluates as a plain
declaration.
Full test:data suite: 5307/5307 pass. typecheck + typecheck:api clean.
* feat(seed-contract): consumer-prices publish.ts writes envelopes
Wrap the 5 canonical keys written by consumer-prices-core/src/jobs/publish.ts
(overview, movers:7d/30d, freshness, categories:7d/30d/90d, retailer-spread,
basket-series) in {_seed, data} envelopes. Legacy seed-meta:<key> writes
preserved for dual-write.
Inlined a buildEnvelope helper (10 lines) rather than taking a cross-package
dependency — consumer-prices-core is a standalone npm package. Documented the
four-file parity contract (mjs source, ts mirror, js edge mirror, this copy).
Contract fields: sourceVersion='consumer-prices-core-publish-v1', schemaVersion=1,
state='OK' (recordCount>0) or 'OK_ZERO' (legitimate zero).
Typecheck: no new errors in publish.ts.
* fix(seed-contract): 3 more server-side readers unwrap envelopes
Found during final audit:
- server/worldmonitor/resilience/v1/_shared.ts: resilience score reader
parsed cached GetResilienceScoreResponse raw. Contract-mode seed-resilience-scores
now envelopes those keys.
- server/worldmonitor/resilience/v1/get-resilience-ranking.ts: p05/p95
interval lookup parsed raw from seed-resilience-scores' extra-key path.
- server/worldmonitor/infrastructure/v1/_shared.ts: mgetJson() used for
count-source keys (wildfire:fires:v1, news:insights:v1) which are both
contract-mode now.
All three now unwrap via server/_shared/seed-envelope. Legacy shapes pass
through unchanged.
Typecheck clean.
* feat(seed-contract): ais-relay.cjs direct writes produce envelopes
32 canonical-key write sites in scripts/ais-relay.cjs now produce {_seed, data}
envelopes. Inlined buildEnvelope() (CJS module can't require ESM source) +
envelopeWrite(key, data, ttlSeconds, meta) wrapper. Enveloped keys span market
bootstrap, aviation, cyber-threats, theater-posture, weather-alerts, economic
spending/fred/worldbank, tech-events, corridor-risk, usni-fleet, shipping-stress,
social:reddit, wsb-tickers, pizzint, product-catalog, chokepoint transits,
ucdp-events, satellites, oref.
Left bare (not seeded data keys): seed-meta:* (dual-write legacy),
classifyCacheKey LLM cache, notam:prev-closed-state internal state,
wm:notif:scan-dedup flags.
Updated tests/ucdp-seed-resilience.test.mjs regex to accept both upstashSet
(pre-contract) and envelopeWrite (post-contract) call patterns.
* feat(seed-contract): 15 bundle files add canonicalKey for envelope gate
54 bundle sections across 12 files now declare canonicalKey alongside the
existing seedMetaKey. _bundle-runner.mjs (from PR 1) prefers canonicalKey
when both are present — gates section runs on envelope._seed.fetchedAt
read directly from the data key, eliminating the meta-outlives-data class
of bugs.
Files touched:
- climate (5), derived-signals (2), ecb-eu (3), energy-sources (6),
health (2), imf-extended (4), macro (10), market-backup (9),
portwatch (4), relay-backup (2), resilience-recovery (5), static-ref (2)
Skipped (14 sections, 3 whole bundles): multi-key writers, dynamic
templated keys (displacement year-scoped), or non-runSeed orchestrators
(regional brief cron, resilience-scores' 222-country publish, validation/
benchmark scripts). These continue to use seedMetaKey or their own gate.
seedMetaKey preserved everywhere — dual-write. _bundle-runner.mjs falls
back to legacy when canonicalKey is absent.
All 15 bundles pass node --check. test:data: 5307/5307. typecheck:all: clean.
* fix(seed-contract): 4 PR #3097 review P1s — transform/declareRecords mismatches + envelope leaks
Addresses both P1 findings and the extra-key seed-meta leak surfaced in review:
1. runSeed helper-level invariant: seed-meta:* keys NEVER envelope.
scripts/_seed-utils.mjs exports shouldEnvelopeKey(key) — returns false for
any key starting with 'seed-meta:'. Both atomicPublish (canonical) and
writeExtraKey (extras) gate the envelope wrap through this helper. Fixes
seed-iea-oil-stocks' ANALYSIS_META_EXTRA_KEY silently getting enveloped,
which broke health.js parsing the value as bare {fetchedAt, recordCount}.
Also defends against any future manual writeExtraKey(..., envelopeMeta)
call that happens to target a seed-meta:* key.
2. seed-token-panels canonical + extras fixed.
publishTransform returns data.defi (the defi panel itself, shape {tokens}).
Old declareRecords counted data.defi.tokens + data.ai.tokens + data.other.tokens
on the transformed payload → 0 → RETRY path → canonical market:defi-tokens:v1
never wrote, and because runSeed returned before the extraKeys loop,
market:ai-tokens:v1 + market:other-tokens:v1 stayed stale too.
New: declareRecords counts data.tokens on the transformed shape. AI_KEY +
OTHER_KEY extras reuse the same function (transforms return structurally
identical panels). Added isMain guard so test imports don't fire runSeed.
3. api/product-catalog.js cached reader unwraps envelope.
ais-relay.cjs now envelopes product-catalog:v2 via envelopeWrite(). The
edge reader did raw JSON.parse(result) and returned {_seed, data} to
clients, breaking the cached path. Fix: import unwrapEnvelope from
./_seed-envelope.js, apply after JSON.parse. One site — :238-241 is
downstream of getFromCache(), so the single reader fix covers both.
4. Regression lock tests/seed-contract-transform-regressions.test.mjs (11 cases):
- shouldEnvelopeKey invariant: seed-meta:* false, canonical true
- Token-panels declareRecords works on transformed shape (canonical + both extras)
- Explicit repro of pre-fix buggy signature returning 0 — guards against revert
- resolveRecordCount accepts 0, rejects non-integer
- Product-catalog envelope unwrap returns bare shape; legacy passes through
Verification:
- npm run test:data → 5318/5318 pass (was 5307 — 11 new regressions)
- npm run typecheck:all → clean
- node --check on every modified script
iea-oil-stocks canonical declareRecords was NOT broken (user confirmed during
review — buildIndex preserves .members); only its ANALYSIS_META_EXTRA_KEY
was affected, now covered generically by commit 1's helper invariant.
* fix(seed-contract): seed-token-panels validateFn also runs on post-transform shape
Review finding: fixing declareRecords wasn't sufficient — atomicPublish() runs
validateFn(publishData) on the transformed payload too. seed-token-panels'
validate() checked data.defi/.ai/.other on the transformed {tokens} shape,
returned false, and runSeed took the early skipped-write branch (before even
reaching the declareRecords RETRY logic). Net effect: same as before the
declareRecords fix — canonical + both extras stayed stale.
Fix: validate() now checks the canonical defi panel directly (Array.isArray
(data?.tokens) && has at least one t.price > 0). AI/OTHER panels are validated
implicitly by their own extraKey declareRecords on write.
Audited the other 9 seeders with publishTransform (bls-series, bis-extended,
bis-data, gdelt-intel, trade-flows, iea-oil-stocks, jodi-gas, sanctions-pressure,
forecasts): all validateFn's correctly target the post-transform shape. Only
token-panels regressed.
Added 4 regression tests (tests/seed-contract-transform-regressions.test.mjs):
- validate accepts transformed panel with priced tokens
- validate rejects all-zero-price tokens
- validate rejects empty/missing tokens
- Explicit pre-fix repro (buggy old signature fails on transformed shape)
Verification:
- npm run test:data → 5322/5322 pass (was 5318; +4 new)
- npm run typecheck:all → clean
- node --check clean
* feat(seed-contract): add /api/seed-contract-probe validation endpoint
Single machine-readable gate for 'is PR #3097 working in production'.
Replaces the curl/jq ritual with one authenticated edge call that returns
HTTP 200 ok:true or 503 + failing check list.
What it validates:
- 8 canonical keys have {_seed, data} envelopes with required data fields
and minRecords floors (fsi-eu, zone-normals, 3 token panels + minRecords
guard against token-panels RETRY regression, product-catalog, wildfire,
earthquakes).
- 2 seed-meta:* keys remain BARE (shouldEnvelopeKey invariant; guards
against iea-oil-stocks ANALYSIS_META_EXTRA_KEY-class regressions).
- /api/product-catalog + /api/bootstrap responses contain no '_seed' leak.
Auth: x-probe-secret header must match RELAY_SHARED_SECRET (reuses existing
Vercel↔Railway internal trust boundary).
Probe logic is exported (checkProbe, checkPublicBoundary, DEFAULT_PROBES) for
hermetic testing. tests/seed-contract-probe.test.mjs covers every branch:
envelope pass/fail on field/records/shape, bare pass/fail on shape/field,
missing/malformed JSON, Redis non-2xx, boundary seed-leak detection,
DEFAULT_PROBES sanity (seed-meta invariant present, token-panels minRecords
guard present).
Usage:
curl -H "x-probe-secret: $RELAY_SHARED_SECRET" \
https://api.worldmonitor.app/api/seed-contract-probe
PR 3 will extend the probe with a stricter mode that asserts seed-meta:*
keys are GONE (not just bare) once legacy dual-write is removed.
Verification:
- tests/seed-contract-probe.test.mjs → 15/15 pass
- npm run test:data → 5338/5338 (was 5322; +16 new incl. conformance)
- npm run typecheck:all → clean
* fix(seed-contract): tighten probe — minRecords on AI/OTHER + cache-path source header
Review P2 findings: the probe's stated guards were weaker than advertised.
1. market:ai-tokens:v1 + market:other-tokens:v1 probes claimed to guard the
token-panels extra-key RETRY regression but only checked shape='envelope'
+ dataHas:['tokens']. If an extra-key declareRecords regressed to 0, both
probes would still pass because checkProbe() only inspects _seed.recordCount
when minRecords is set. Now both enforce minRecords: 1.
2. /api/product-catalog boundary check only asserted no '_seed' leak — which
is also true for the static fallback path. A broken cached reader
(getFromCache returning null or throwing) could serve fallback silently
and still pass this probe. Now:
- api/product-catalog.js emits X-Product-Catalog-Source: cache|dodo|fallback
on the response (the json() helper gained an optional source param wired
to each of the three branches).
- checkPublicBoundary declaratively requires that header's value match
'cache' for /api/product-catalog, so a fallback-serve fails the probe
with reason 'source:fallback!=cache' or 'source:missing!=cache'.
Test updates (tests/seed-contract-probe.test.mjs):
- Boundary check reworked to use a BOUNDARY_CHECKS config with optional
requireSourceHeader per endpoint.
- New cases: served-from-cache passes, served-from-fallback fails with source
mismatch, missing header fails, seed-leak still takes precedence, bad
status fails.
- Token-panels sanity test now asserts minRecords≥1 on all 3 panels.
Verification:
- tests/seed-contract-probe.test.mjs → 17/17 pass (was 15, +2 net)
- npm run test:data → 5340/5340
- npm run typecheck:all → clean
250 lines
9.7 KiB
JavaScript
250 lines
9.7 KiB
JavaScript
#!/usr/bin/env node
|
||
// Seed UN Comtrade strategic commodity trade flows (issue #2045).
|
||
// Uses the public preview endpoint — no auth required.
|
||
|
||
import { loadEnvFile, CHROME_UA, runSeed, sleep, writeExtraKey } from './_seed-utils.mjs';
|
||
|
||
loadEnvFile(import.meta.url);
|
||
|
||
const CANONICAL_KEY = 'comtrade:flows:v1';
|
||
const CACHE_TTL = 259200; // 72h = 3× daily interval
|
||
export const KEY_PREFIX = 'comtrade:flows';
|
||
const COMTRADE_BASE = 'https://comtradeapi.un.org/public/v1';
|
||
const INTER_REQUEST_DELAY_MS = 3_000;
|
||
const ANOMALY_THRESHOLD = 0.30; // 30% YoY change
|
||
// Require at least this fraction of (reporter × commodity) pairs to return
|
||
// non-empty flows. Guards against an entire reporter silently flatlining
|
||
// (e.g., wrong reporterCode → HTTP 200 with count:0 for every commodity).
|
||
// Global coverage floor — overall populated/total must be ≥ this.
|
||
const MIN_COVERAGE_RATIO = 0.70;
|
||
// Per-reporter coverage floor — each reporter must have ≥ this fraction of
|
||
// its commodities populated. Prevents the "India/Taiwan flatlines entirely"
|
||
// failure mode: with 6 reporters × 5 commodities, losing one full reporter
|
||
// is only 5/30 missing (83% global coverage → passes MIN_COVERAGE_RATIO),
|
||
// but 0/5 per-reporter coverage for the dead one blocks publish here.
|
||
const MIN_PER_REPORTER_RATIO = 0.40; // at least 2 of 5 commodities per reporter
|
||
|
||
// Strategic reporters: US, China, Russia, Iran, India, Taiwan
|
||
const REPORTERS = [
|
||
{ code: '842', name: 'USA' },
|
||
{ code: '156', name: 'China' },
|
||
{ code: '643', name: 'Russia' },
|
||
{ code: '364', name: 'Iran' },
|
||
{ code: '699', name: 'India' },
|
||
{ code: '490', name: 'Taiwan' },
|
||
];
|
||
|
||
// Strategic HS commodity codes
|
||
const COMMODITIES = [
|
||
{ code: '2709', desc: 'Crude oil' },
|
||
{ code: '2711', desc: 'LNG / natural gas' },
|
||
{ code: '7108', desc: 'Gold' },
|
||
{ code: '8542', desc: 'Semiconductors' },
|
||
{ code: '9301', desc: 'Arms / military equipment' },
|
||
];
|
||
|
||
// Comtrade preview regularly hits transient 5xx (500/502/503/504). Without
|
||
// retry each (reporter,commodity) pair that drew a 5xx is silently lost.
|
||
export function isTransientComtrade(status) {
|
||
return status === 500 || status === 502 || status === 503 || status === 504;
|
||
}
|
||
|
||
// Injectable sleep so unit tests can exercise the retry loop without real
|
||
// 5s/15s waits. Production defaults to the real sleep.
|
||
let _retrySleep = sleep;
|
||
export function __setSleepForTests(fn) { _retrySleep = typeof fn === 'function' ? fn : sleep; }
|
||
|
||
export async function fetchFlows(reporter, commodity) {
|
||
const url = new URL(`${COMTRADE_BASE}/preview/C/A/HS`);
|
||
url.searchParams.set('reporterCode', reporter.code);
|
||
url.searchParams.set('cmdCode', commodity.code);
|
||
url.searchParams.set('flowCode', 'X,M'); // exports + imports
|
||
|
||
async function once() {
|
||
return fetch(url.toString(), {
|
||
headers: { 'User-Agent': CHROME_UA, Accept: 'application/json' },
|
||
signal: AbortSignal.timeout(15_000),
|
||
});
|
||
}
|
||
|
||
// Classification loop: up to two transient-5xx retries (5s, 15s) then give up.
|
||
let transientRetries = 0;
|
||
const MAX_TRANSIENT_RETRIES = 2;
|
||
let resp;
|
||
while (true) {
|
||
resp = await once();
|
||
if (isTransientComtrade(resp.status) && transientRetries < MAX_TRANSIENT_RETRIES) {
|
||
const delay = transientRetries === 0 ? 5_000 : 15_000;
|
||
console.warn(` transient HTTP ${resp.status} for reporter ${reporter.code} cmd ${commodity.code}, retrying in ${delay / 1000}s...`);
|
||
await _retrySleep(delay);
|
||
transientRetries++;
|
||
continue;
|
||
}
|
||
break;
|
||
}
|
||
|
||
if (!resp.ok) throw new Error(`HTTP ${resp.status}`);
|
||
const data = await resp.json();
|
||
|
||
// Comtrade preview returns { data: [...] } with annual records
|
||
const records = data?.data ?? [];
|
||
if (!Array.isArray(records)) return [];
|
||
|
||
// The preview endpoint returns partner-level rows (one per counterparty).
|
||
// Aggregate to World totals per (flowCode, year) by summing, so YoY is
|
||
// computed against full-year totals. Keying on (flowCode, year) without
|
||
// summing would silently drop every partner except the last one seen.
|
||
const byFlowYear = new Map(); // key: `${flowCode}:${year}`
|
||
for (const r of records) {
|
||
const year = Number(r.period ?? r.refYear ?? r.refMonth?.slice(0, 4) ?? 0);
|
||
if (!year) continue;
|
||
const flowCode = String(r.flowCode ?? r.rgDesc ?? 'X');
|
||
const val = Number(r.primaryValue ?? r.cifvalue ?? r.fobvalue ?? 0);
|
||
const wt = Number(r.netWgt ?? 0);
|
||
const mapKey = `${flowCode}:${year}`;
|
||
const prev = byFlowYear.get(mapKey);
|
||
if (prev) {
|
||
prev.val += val;
|
||
prev.wt += wt;
|
||
} else {
|
||
byFlowYear.set(mapKey, { year, flowCode, val, wt, partnerCode: '000', partnerName: 'World' });
|
||
}
|
||
}
|
||
|
||
// Derive the set of (flowCode, year) pairs sorted for YoY lookup.
|
||
const entries = Array.from(byFlowYear.values()).sort((a, b) => a.year - b.year || a.flowCode.localeCompare(b.flowCode));
|
||
const flows = [];
|
||
|
||
for (const cur of entries) {
|
||
const prevKey = `${cur.flowCode}:${cur.year - 1}`;
|
||
const prev = byFlowYear.get(prevKey);
|
||
const yoyChange = prev && prev.val > 0 ? (cur.val - prev.val) / prev.val : 0;
|
||
const isAnomaly = Math.abs(yoyChange) > ANOMALY_THRESHOLD;
|
||
|
||
flows.push({
|
||
reporterCode: reporter.code,
|
||
reporterName: reporter.name,
|
||
partnerCode: cur.partnerCode,
|
||
partnerName: cur.partnerName,
|
||
cmdCode: commodity.code,
|
||
cmdDesc: commodity.desc,
|
||
year: cur.year,
|
||
tradeValueUsd: cur.val,
|
||
netWeightKg: cur.wt,
|
||
yoyChange,
|
||
isAnomaly,
|
||
});
|
||
}
|
||
|
||
return flows;
|
||
}
|
||
|
||
async function fetchAllFlows() {
|
||
const allFlows = [];
|
||
const perKeyFlows = {};
|
||
|
||
for (let ri = 0; ri < REPORTERS.length; ri++) {
|
||
for (let ci = 0; ci < COMMODITIES.length; ci++) {
|
||
const reporter = REPORTERS[ri];
|
||
const commodity = COMMODITIES[ci];
|
||
const label = `${reporter.name}/${commodity.desc}`;
|
||
|
||
if (ri > 0 || ci > 0) await sleep(INTER_REQUEST_DELAY_MS);
|
||
console.log(` Fetching ${label}...`);
|
||
|
||
let flows = [];
|
||
try {
|
||
flows = await fetchFlows(reporter, commodity);
|
||
console.log(` ${flows.length} records`);
|
||
} catch (err) {
|
||
console.warn(` ${label}: failed (${err.message})`);
|
||
}
|
||
|
||
allFlows.push(...flows);
|
||
const key = `${KEY_PREFIX}:${reporter.code}:${commodity.code}`;
|
||
perKeyFlows[key] = { flows, fetchedAt: new Date().toISOString() };
|
||
}
|
||
}
|
||
|
||
const gate = checkCoverage(perKeyFlows, REPORTERS, COMMODITIES);
|
||
console.log(` Coverage: ${gate.populated}/${gate.total} (${(gate.globalRatio * 100).toFixed(0)}%) reporter×commodity pairs populated`);
|
||
for (const r of gate.perReporter) {
|
||
if (r.ratio < MIN_PER_REPORTER_RATIO) {
|
||
console.warn(` ${r.reporter} reporter ${r.code}: ${r.populated}/${r.total} (${(r.ratio * 100).toFixed(0)}%) — below per-reporter floor ${MIN_PER_REPORTER_RATIO}`);
|
||
}
|
||
}
|
||
if (!gate.ok) throw new Error(gate.reason);
|
||
|
||
return { flows: allFlows, perKeyFlows, fetchedAt: new Date().toISOString() };
|
||
}
|
||
|
||
/**
|
||
* Pure coverage gate. Returns pass/fail + per-reporter breakdown.
|
||
* Exported for unit testing — mocking 30+ fetches in fetchAllFlows is fragile,
|
||
* and the failure mode the PR is trying to block lives here, not in fetchFlows.
|
||
*
|
||
* Blocks publish when EITHER: global ratio < MIN_COVERAGE_RATIO, OR any single
|
||
* reporter's commodity coverage < MIN_PER_REPORTER_RATIO. The latter catches
|
||
* the India/Taiwan-style "one reporter flatlines completely" case that passes
|
||
* a global-only gate.
|
||
*/
|
||
export function checkCoverage(perKeyFlows, reporters, commodities) {
|
||
const total = reporters.length * commodities.length;
|
||
const populated = Object.values(perKeyFlows).filter((v) => (v.flows?.length ?? 0) > 0).length;
|
||
const globalRatio = total > 0 ? populated / total : 0;
|
||
|
||
const perReporter = reporters.map((r) => {
|
||
const pop = commodities.filter((c) => (perKeyFlows[`${KEY_PREFIX}:${r.code}:${c.code}`]?.flows?.length ?? 0) > 0).length;
|
||
return { reporter: r.name, code: r.code, populated: pop, total: commodities.length, ratio: commodities.length > 0 ? pop / commodities.length : 0 };
|
||
});
|
||
|
||
if (globalRatio < MIN_COVERAGE_RATIO) {
|
||
return { ok: false, populated, total, globalRatio, perReporter, reason: `coverage ${populated}/${total} below global floor ${MIN_COVERAGE_RATIO}; refusing to publish partial snapshot` };
|
||
}
|
||
const dead = perReporter.find((r) => r.ratio < MIN_PER_REPORTER_RATIO);
|
||
if (dead) {
|
||
return { ok: false, populated, total, globalRatio, perReporter, reason: `reporter ${dead.reporter} (${dead.code}) only ${dead.populated}/${dead.total} commodities — below per-reporter floor ${MIN_PER_REPORTER_RATIO}; refusing to publish snapshot with a flatlined reporter` };
|
||
}
|
||
return { ok: true, populated, total, globalRatio, perReporter, reason: null };
|
||
}
|
||
|
||
function validate(data) {
|
||
return Array.isArray(data?.flows) && data.flows.length > 0;
|
||
}
|
||
|
||
function publishTransform(data) {
|
||
const { perKeyFlows: _pkf, ...rest } = data;
|
||
return rest;
|
||
}
|
||
|
||
async function afterPublish(data, _meta) {
|
||
for (const [key, value] of Object.entries(data.perKeyFlows ?? {})) {
|
||
if ((value.flows?.length ?? 0) > 0) {
|
||
await writeExtraKey(key, value, CACHE_TTL);
|
||
}
|
||
}
|
||
}
|
||
|
||
// isMain guard so tests can import fetchFlows without triggering a real seed run.
|
||
export function declareRecords(data) {
|
||
return Array.isArray(data?.flows) ? data.flows.length : 0;
|
||
}
|
||
|
||
if (process.argv[1]?.endsWith('seed-trade-flows.mjs')) {
|
||
runSeed('trade', 'comtrade-flows', CANONICAL_KEY, fetchAllFlows, {
|
||
validateFn: validate,
|
||
ttlSeconds: CACHE_TTL,
|
||
sourceVersion: 'comtrade-preview-v1',
|
||
publishTransform,
|
||
afterPublish,
|
||
|
||
declareRecords,
|
||
schemaVersion: 1,
|
||
maxStaleMin: 2880,
|
||
}).catch((err) => {
|
||
const _cause = err.cause ? ` (cause: ${err.cause.message || err.cause.code || err.cause})` : '';
|
||
console.error('FATAL:', (err.message || err) + _cause);
|
||
process.exit(0);
|
||
});
|
||
}
|