Latency-Distribution-Aware Child-Order Microbatch Slippage Playbook

2026-03-11 · finance

Latency-Distribution-Aware Child-Order Microbatch Slippage Playbook

Date: 2026-03-11
Category: research
Scope: Live execution where wire/ACK jitter is material (equities/futures, venue-agnostic)


1) Why this topic matters

Most schedulers assume each child order is emitted instantly and independently.

In production, that is false:

Result: a "smooth" schedule on paper can become a toxic burst schedule on the tape, increasing slippage tails.

The practical fix is not just "send slower". It is to optimize microbatch size and spacing using the observed latency distribution and alpha half-life.


2) Core idea

At each decision time, choose among emission patterns:

Use a model that prices both:

  1. impact from immediate footprint,
  2. stale-arrival risk from latency distribution.

3) Minimal model

Let:

3.1 Stale-arrival probability

If edge decays exponentially, stale risk for one slice:

[ p_{stale} = \mathbb{P}(L > h) = 1 - F_L(h) ]

where (F_L) is empirical CDF of latency for this route/venue/time bucket.

3.2 Cost decomposition per slice (buy side)

[ C = C_{spread} + C_{impact} + C_{stale} + C_{underfill} ]

Simple operational parameterization:

[ C_{spread} \approx \frac{1}{2},spread \cdot \mathbb{1}_{aggr} ]

[ C_{impact} \approx a\left(\frac{x}{D}\right)^\beta ]

[ C_{stale} \approx b,\sigma,\sqrt{\mathbb{E}[L]},p_{stale} ]

[ C_{underfill} \approx c,\text{deadline gap penalty} ]

Batch objective (tail-aware):

[ \min_{k,\Delta,x} ; \mathbb{E}[C_{batch}] + \lambda,CVaR_{95}(C_{batch}) ]

subject to participation, risk, and completion constraints.


4) Why microbatching helps (and when it hurts)

Helps when

Hurts when

Rule of thumb: if (p_{stale}) jumps while depth is thinning, microbatch should shrink to SINGLE or controlled aggression.


5) Feature set for a production model

Model outputs per candidate action:


6) Online controller (practical policy)

Every control step (e.g., 100-300ms):

  1. Build candidate actions: SINGLE, MICROBATCH(2..5, Δ∈{20,40,80}ms), HOLD.
  2. Score each candidate with expected + tail objective.
  3. Apply hard guards:
    • if reject/backlog high → cap k and disable tight Δ,
    • if deadline near → disallow HOLD,
    • if toxicity spike → reduce passive dwell and batch size.
  4. Execute best admissible action.

Emergency downgrade path:


7) Calibration & monitoring

Weekly calibration

Intraday monitors

Auto fallback trigger examples:


8) Backtest / replay design

Use event replay with injected empirical latency traces (not fixed latency):

A/B/C policies:

Report by regime:

Promotion criterion: tail reduction with no material completion deterioration.


9) Implementation checklist


10) References


One-line takeaway

If latency is random and heavy-tailed, execution should optimize when and how many child orders to emit per burst—not just how much to trade per minute.