Replication and Lab Limits
TLDR: Treat lab effects as hypotheses, and use field validation to decide. Prefer effects with strong real‑world support. Beware of headline claims without denominators and windows.
Replication context
- Social/behavioral sciences show mixed replication rates; many lab findings fail under natural contexts and incentives.
- Loss aversion is context‑dependent: robust for large losses; broad “losses loom larger” claims are contested.
- Publication bias and p‑hacking inflate apparent effects; practitioners must validate locally.
Practitioner guidance
- Start from behaviors and contexts, not a catalog of “effects.”
- Run small field pilots with Δ‑B and TTFB instrumentation before scale.
- Treat “defaults,” “priming,” and “magic words” with skepticism unless proven in your segment/context.
- Use Evidence Ledger entries to ground claims and limit overreach (see BS‑0003, BS‑0004).
Why lab → field often breaks
- Context shifts: stakes, social setting, timing, and device all change behavior.
- Interaction effects: multiple components interact (ability × motivation × environment); single‑factor lab manipulations rarely hold alone.
- Measurement drift: proxies (clicks, toggles) do not equal target behaviors.
What travels well
- Fit‑first method: validate problem seeking, then behavior feasibility/desirability.
- Friction removal: TTFB reductions for the same behavior in the same context.
- Clear value loops: immediate reinforcement for first successful behavior.
See also: Why Nudges Fail, Nudge Limitations, Measurement Standards.