How It Works

End-to-end runtime behavior, core concepts, and pipeline walkthrough.

This guide explains dagzoo end-to-end with enough detail to reason about behavior without needing to jump between many documents.

Who this is for

  • End users running dagzoo generate and dagzoo benchmark
  • Users who want a deeper mental model of runtime behavior and outputs

Mental model in 90 seconds

dagzoo synthesizes tabular datasets by sampling causal structure, executing randomized mechanisms over that structure, and enforcing quality and realism controls.

  1. Resolve config and hardware context for the command.
  2. Derive deterministic seeds for run, dataset, and component scopes.
  3. Sample a dataset layout (feature types, assignments, graph size bounds).
  4. Sample a latent DAG plus feature-to-node and target-to-node assignments.
  5. Execute node pipelines in topological order to produce latent feature outputs.
  6. Convert latent feature outputs into complete X_complete.
  7. Convert one selected latent node into the emitted target y.
  8. Apply split checks and target postprocess, then optionally apply missingness as a joint observation process before emission.
  9. Emit DatasetBundle outputs; optionally persist shards and diagnostics.
  10. Persist shard outputs and effective-config artifacts; when needed, run dagzoo filter later as a separate post-generation stage.

Core Concepts

1. Causal DAG vs tabular layout

The generation graph is a latent DAG, while emitted columns are a tabular projection of that latent graph.

  • Latent nodes represent abstract causal variables.
  • Feature columns are assigned to latent nodes by sampled layout state.
  • The target is assigned to one latent node and emitted from that node’s converter stack.
  • Multiple columns can map to one node, and one node can influence many columns.
  • This decoupling allows rich causal interactions while preserving a clean acyclic execution graph.
  • This is also important for prior diversity: a single latent DAG can produce many different observed tabular layouts (different feature-to-node assignments, different converter choices), so effective diversity grows faster than the number of unique DAG topologies sampled.
flowchart LR
    %% Class Definitions
    classDef latent fill:#e1f5fe,stroke:#01579b,stroke-width:2px,color:#01579b,stroke-dasharray: 5 5
    classDef observable fill:#f5f5f5,stroke:#212121,stroke-width:2px,color:#212121

    subgraph LatentSpace [Latent Causal DAG]
        A((Node A)) --> B((Node B))
        A --> C((Node C))
    end

    subgraph ObservableSpace [Tabular Layout]
        F1[feature_0 num]
        F2[feature_1 cat]
        F3[feature_2 num]
        T[target]
    end

    %% Mapping connections
    A -. mapping .-> F1
    A -. mapping .-> F2
    B -. mapping .-> F3
    C -. target mapping .-> T

    %% Assign Classes
    class A,B,C latent
    class F1,F2,F3,T observable

    style LatentSpace fill:#f0faff,stroke:#01579b,stroke-dasharray: 5 5
    style ObservableSpace fill:#fafafa,stroke:#212121

2. Reproducibility tree

Reproducibility is driven by KeyedRng, which maps one base seed onto named semantic subtrees.

  • One run seed fans out into deterministic run, dataset, layout, split, missingness, noise, and benchmark subtrees.
  • Public metadata exposes stable replay identifiers such as seed, dataset_index, run_num_datasets, and dataset_seed.
  • Some canonical outputs also include detailed replay metadata such as metadata.keyed_replay when exact run reconstruction matters.
  • Changing one component path should not perturb unrelated component randomness.

Illustrative derivation chain:

KeyedRng(run_seed)
  -> keyed("rows")
  -> keyed("plan_candidate", attempt, "layout")
  -> keyed("plan_candidate", attempt, "execution_plan")
  -> keyed("dataset", i, "noise_runtime")
  -> keyed("dataset", i, "attempt", attempt, "split")
  -> keyed("dataset", i, "attempt", attempt, "postprocess")
  -> keyed("dataset", i, "attempt", attempt, "missingness")

3. Split validity retries and post-generation filtering

Generation retries cover split-validity and generation exceptions only.

  • Retries are bounded by filter.max_attempts (retry budget reused from config), and generation reuses filter.min_target_* thresholds when it resamples layouts for structural validity regardless of filter.enabled.
  • Emitted metadata records attempt_used and generation-attempt counters.
  • Generated outputs record metadata.filter.mode=deferred and metadata.filter.status=not_run so later filter runs can be traced back to the original generated corpus.

Acceptance is a separate post-generation stage:

  • dagzoo filter replays structural lineage-validity checks over emitted shards and writes accepted/rejected outcomes after generation.
  • Request-driven handoff currently publishes generated shards only.

4. Effective config and traceability

Generation and benchmark commands resolve effective config through staged overrides, then validate constraints.

Generate path (high-level):

  1. Base config YAML
  2. Device override normalization
  3. Hardware policy application
  4. Missingness/diagnostics CLI overrides
  5. Final generation constraint validation

Every run writes:

  • effective_config.yaml
  • effective_config_trace.yaml

The trace is field-level provenance (path, source, old_value, new_value) for resolved settings.

5. Hardware-aware execution semantics

dagzoo tracks three related but distinct runtime notions:

  • requested_device: normalized user intent (auto, cpu, cuda, mps)
  • resolved_device: backend selected from request/environment
  • device: backend used for dataset execution in that attempt

Notable runtime behavior:

  • auto resolves to available accelerator first, else CPU.
  • Backend runtime errors surface directly; generation does not rewrite the resolved device after execution starts.
  • Split/postprocess control RNG runs on CPU to avoid tiny-op accelerator overhead.

Mathematical Foundations

The generation pipeline is a multi-axis sampling process. Each formal section in transforms.md corresponds to an independent axis of prior diversity:

graph structure  ──┐
shift parameters ──┤
mechanism families ┤── together determine the region of meta-feature
node pipeline    ──┤   space the corpus covers (= effective diversity)
converters       ──┤
noise families   ──┘

Broadening one axis (e.g., adding noise families) increases diversity along that dimension without affecting the others — the axes are designed to be independently controllable. The formal specification matters because the parameterization determines what prior regions are reachable: if a transform’s math restricts certain behaviors, no config change can produce datasets in those regions.

Quick index to the formal sections:

  1. DAG sampling: strict upper-triangular Bernoulli sampling with Cauchy latent logits and shift-adjusted edge bias.
  2. Mechanism-family sampling: family-mix weights plus mechanism logit tilt produce runtime family probabilities.
  3. Node pipeline: root/parent composition, latent sanitization and weighting, converter slicing, and final scaling.
  4. Converters and noise: numeric/categorical converter equations and dataset-level noise runtime selection (including mixture-mode behavior).

End-to-end flow

This diagram shows command-level orchestration and where generation, benchmarking, and hardware inspection diverge.

flowchart TB
    %% Class Definitions
    classDef cli fill:#e1f5fe,stroke:#01579b,stroke-width:2px,color:#01579b
    classDef gen fill:#fff3e0,stroke:#e65100,stroke-width:2px,color:#e65100
    classDef bench fill:#f3e5f5,stroke:#4a148c,stroke-width:2px,color:#4a148c
    classDef flt fill:#e8f5e9,stroke:#1b5e20,stroke-width:2px,color:#1b5e20
    classDef hw fill:#f1f8e9,stroke:#33691e,stroke-width:2px,color:#33691e

    CLI(["dagzoo CLI"])

    CLI --> GenCfg
    CLI --> BenchCfg
    CLI --> Detect

    subgraph GeneratePath [generate]
        direction TB
        subgraph Setup [" "]
            GenCfg[load + resolve config] --> Loop[generate_batch_iter]
            Loop --> Seed[derive dataset/component seeds]
        end
        subgraph Sampling [" "]
            Seed --> Layout[sample layout]
            Layout --> DAG[sample DAG + assignments]
            DAG --> Resolve[resolve shift + noise]
        end
        subgraph Execution [" "]
            Resolve --> Exec[run node pipelines]
        end
        subgraph Emission [" "]
            Exec --> Post[postprocess]
            Post --> Missing[optional missingness]
            Missing --> Bundle[[emit DatasetBundle]]
        end
    end

    subgraph FilterPath [filter]
        direction TB
        FilterCmd[dagzoo filter] --> Replay[replay structural lineage checks over shards]
        Replay --> FilterArtifacts[write manifest + summary + optional curated shards]
    end
    Bundle --> FilterCmd

    subgraph BenchmarkPath [benchmark]
        direction TB
        BenchCfg[resolve preset/suite configs] --> Runs[run benchmark suite]
        Runs --> Guards[emit guardrails]
        Guards --> Reports[write summary]
    end

    subgraph HardwarePath [hardware]
        direction TB
        Detect[detect_hardware] --> Print[print backend/tier]
    end

    %% Assign Classes
    class CLI cli
    class GenCfg,Loop,Seed,Layout,DAG,Resolve,Exec,Post,Missing,Bundle gen
    class FilterCmd,Replay,FilterArtifacts flt
    class BenchCfg,Runs,Guards,Reports bench
    class Detect,Print hw

    style Setup fill:transparent,stroke:none
    style Sampling fill:transparent,stroke:none
    style Execution fill:transparent,stroke:none
    style Emission fill:transparent,stroke:none

Generation pipeline walkthrough

This section maps the runtime to the main execution phases and data flow.

1) Entry points and orchestration boundaries

  • dagzoo generate and build_dataloader(...) use the same public generation flow.
  • Generation resolves config and hardware context, derives deterministic seeds, samples layout and structure, executes the latent graph, then finalizes outputs and metadata.
  • dagzoo filter is a later acceptance/replay stage over emitted shards.
  • dagzoo benchmark runs preset or custom suites over the same public generation surface.

2) Layout and structure sampling

  • _sample_layout samples feature counts/types, class bounds, and feature-to-node assignments.
  • sample_dag samples strict upper-triangular adjacency.
  • Adjacency convention is adjacency[src, dst]; parents of node j are read from column adjacency[:, j].

3) Node execution and tensor assembly

  • Nodes execute in index/topological order.
  • Root nodes sample latent points directly.
  • Child nodes consume parent outputs using multi-parent composition:
    • 50% path: concatenate parents and apply one mechanism
    • 50% path: apply per-parent mechanisms, then aggregate via sum | product | max | logsumexp
  • Converter specs slice latent columns and emit feature values.
  • Unassigned feature slots are filled with sampled noise.
  • The selected target node emits raw targets through its target converter.

4) Quality, shift/noise controls, and postprocessing

  • Shift runtime params modulate graph/mechanism/noise behavior when enabled.
  • Noise runtime resolution picks one family per dataset in mixture mode, then propagates through node-level samplers.
  • Split, target postprocess, and optional missingness run in-generation.
  • Classification split validity is enforced before bundle emission.

Current public prior behavior:

  • The shipped generator samples a latent DAG, emits features from node-assigned converters, and emits the target from one selected latent node.
  • Optional missingness is applied afterward as an observation process over the emitted feature table.

Current public postprocess behavior:

  • Public generation defaults to heterogeneous per-dataset layouts, so emitted schema may vary within one run.
  • Stratified mode batches compatible heterogeneous strata without forcing one emitted schema across the run.
  • Numeric feature clipping and standardization are fit on the emitted training split, then applied unchanged to the emitted test split; regression targets follow the same train-fit postprocess rule.
  • Classification runs emit bundles as they are finalized; a later dataset can still exhaust the retry budget after earlier bundles have already been emitted.

5) Metadata and output emission

Each bundle includes runtime metadata for lineage, filter status, shift, noise distribution, and resolved config snapshot.

  • lineage aligns emitted feature columns with DAG node assignments and records the selected latent target node plus emitted target-relevance summary.
  • requested_device, resolved_device, and the reserved device_fallback_reason field are emitted for runtime observability.
  • Public generation outputs add layout_mode, layout_plan_seed, layout_signature, dataset_seed, and keyed_replay.

DAG/node data flow

This diagram focuses on node-level execution mechanics inside the shared generation runtime.

flowchart TB
    %% Class Definitions
    classDef setup fill:#e1f5fe,stroke:#01579b,stroke-width:2px,color:#01579b
    classDef loop fill:#fff3e0,stroke:#e65100,stroke-width:2px,color:#e65100
    classDef pipe fill:#f3e5f5,stroke:#4a148c,stroke-width:2px,color:#4a148c
    classDef out fill:#f1f8e9,stroke:#33691e,stroke-width:2px,color:#33691e

    Setup[layout + DAG + feature node specs] --> Walk[for node in topological order]
    Walk --> Root{root node?}

    Root -- yes --> RootSample[sample_random_points]
    Root -- no --> ParentMix[compose parent outputs]

    RootSample --> Pipe
    ParentMix --> Pipe

    subgraph Pipe [apply_node_pipeline]
        Clean[nan_to_num + clamp] --> Std[standardize]
        Std --> Weight[random weights + normalize]
        Weight --> Slice[slice latent per spec]
        Slice --> Convert[apply converter]
        Convert --> Feedback[write output back]
        Feedback --> Scale[final scaling]
    end

    Pipe --> Emit[emit node output + extracted values]
    Emit --> Walk
    Emit --> Assemble[assemble raw feature matrix]
    Assemble --> PostX[feature postprocess]
    PostX --> TargetNode[convert selected target node]
    TargetNode --> Missing[apply optional observation missingness]
    Missing --> Out[[return emitted X + raw y + filter status metadata]]

    %% Assign Classes
    class Setup setup
    class Walk,Root,RootSample,ParentMix loop
    class Clean,Std,Weight,Slice,Convert,Feedback,Scale,Pipe pipe
    class Emit,Assemble,Out out

Diagnostics, heterogeneous generation, and benchmark guardrails

These are related but distinct runtime surfaces.

  • Public generation now defaults to heterogeneous per-dataset layout and plan sampling.
  • Stratified mode still keeps per-dataset layout semantics, but batches compatible exact strata for throughput-sensitive large runs.
  • Diagnostics aggregates observability metrics across emitted bundles.
  • Benchmark guardrails evaluate runtime/metadata regressions in suite runs.

Glossary quick reference

  • layout: sampled feature/task scaffold for one dataset.
  • DAG adjacency: upper-triangular parent-child matrix, src -> dst.
  • node pipeline: per-node transform and converter execution path.
  • converter spec: instruction for extracting observable feature slices.
  • target node: the latent DAG node selected to emit the raw target column.
  • filter stage: optional post-generation gate that replays lineage reachability and target-validity checks over emitted shards.
  • filter rejection reason: explicit structural reject code such as target_root or no_feature_target_path recorded in filter manifests and summaries.
  • shift runtime params: resolved graph/mechanism/noise drift controls.
  • noise runtime selection: per-dataset resolved noise family/params.
  • execution plan: resolved layout-and-execution payload prepared for one dataset or one compatible batched stratum.
  • layout signature: deterministic hash fingerprint of a sampled layout.
  • DatasetBundle: in-memory output container with tensors + metadata.
  • effective config trace: field-level override provenance artifact.

Where to go next

  • Canonical transform equations and symbol definitions: docs/transforms.md
  • Output schema and metadata contract: docs/output-format.md
  • CLI workflow examples: docs/usage-guide.md
  • Recipe catalog and citations: docs/reference-packs.md