The hardware and bandwidth for this mirror is donated by METANET, the Webhosting and Full Service-Cloud Provider.
If you wish to report a bug, or if you are interested in having us mirror your free-software or open-source project, please feel free to contact us at mirror[@]metanet.ch.

Sensitivity Analysis: Deficiency vs. E-values

Deniz Akdemir

2026-03-26

Introduction

Sensitivity analysis answers: “How strong would unmeasured confounding need to be to change our conclusions?”

Several frameworks exist:

Framework Key Metric Interpretation
E-value Risk ratio needed “Unmeasured confounder must have RR = X”
Partial R² Variance explained “Confounder must explain X% of variance”
Le Cam Deficiency (δ) Information loss “Transfer penalty bounded by \(M\delta\)

This vignette shows how deficiency-based sensitivity analysis relates to and extends traditional approaches.


Conceptual Translation

The E-value Perspective

The E-value (VanderWeele & Ding, 2017) asks:

“To explain away the observed effect, an unmeasured confounder would need to have a risk ratio of at least E with both treatment and outcome.”

For an observed risk ratio RR, the E-value is:

\[E = RR + \sqrt{RR \times (RR - 1)}\]

The Deficiency Perspective

Deficiency (δ) takes a decision-theoretic view:

“Given the information gap between observational and interventional data, the worst-case regret inflation term is bounded by \(M\delta\) (and there is a minimax floor \((M/2)\delta\)).”

Key insight: δ directly quantifies policy consequences, not just statistical associations.

Conceptual Mapping

E-value Concept Deficiency Equivalent
“Effect explained away” δ → 1 (maximal deficiency)
“Effect robust” δ → 0 (zero deficiency)
E-value = 2 Moderate unmeasured confounding
E-value = 5 Strong unmeasured confounding

Practical Example

Setup

Deficiency Estimation

Confounding Frontier

The confounding frontier maps deficiency as a function of confounding strength:

Reading the Plot:

Policy Regret Bound


Comparison with E-values

Computing an Approximate E-value

For comparison, we can compute the E-value for our effect estimate:

Deficiency vs. E-value: Key Differences

Aspect E-value Deficiency (δ)
Scale Risk ratio Total variation distance
Interpretation Strength needed to “explain away” Information loss for decisions
Decision utility Abstract Direct (via \(M\delta\) transfer penalty)
Multi-method Single estimate Compares strategies
Negative controls Not integrated Built-in diagnostics

When to Use Each

Use E-values when: - Communicating to epidemiologists/clinicians familiar with RR - Binary outcomes with clear risk ratio interpretation - Want a single summary number

Use Deficiency (δ) when: - Need decision-theoretic bounds (policy regret) - Comparing multiple adjustment strategies - Have negative control outcomes available - Working with non-binary outcomes (continuous, survival) - Need to combine with sensitivity frontiers


Extended Sensitivity Analysis

Benchmarking Observed Covariates

A key advantage of the frontier approach is benchmarking: we can see where observed covariates fall on the confounding map.

Using Benchmarks:

The benchmarks show the inferred confounding strength of each observed covariate. If an unmeasured confounder would need to be “stronger than W3” (which we know explains 81% of U’s variance), conclusions are robust.

Combining with Negative Controls

# Add negative control
df$Y_nc <- U + rnorm(n, sd = 0.5) # Affected by U, not by A

spec_full <- causal_spec(
    df, "A", "Y", c("W", "W2", "W3"),
    negative_control = "Y_nc"
)
#> ✔ Created causal specification: n=500, 3 covariate(s)

# Complete analysis
def_full <- estimate_deficiency(
    spec_full,
    methods = c("unadjusted", "iptw"),
    n_boot = 100
)
#> ℹ Estimating deficiency: unadjusted
#> ℹ Estimating deficiency: iptw

nc_full <- nc_diagnostic(spec_full, method = "iptw", n_boot = 100)
#> ℹ Using kappa = 1 (conservative). Consider domain-specific estimation or sensitivity analysis via kappa_range.
#> ✔ No evidence against causal assumptions (p = 0.74257 )

print(def_full)
#> 
#> -- Deficiency Proxy Estimates (PS-TV) ------
#> 
#>      Method  Delta     SE               CI            Quality
#>  unadjusted 0.1033 0.0233 [0.1016, 0.1825] Insufficient (Red)
#>        iptw 0.0204 0.0089  [0.016, 0.0482]  Excellent (Green)
#> Note: delta is a propensity-score TV proxy (overlap/balance diagnostic).
#> 
#> Best method: iptw (delta = 0.0204 )
print(nc_full)
#> 
#> -- Negative Control Diagnostic ----------------------------------------
#> 
#> * screening statistic (weighted corr): 0.0173 
#> * delta_NC (association proxy): 0.0173 
#> * delta bound (under kappa alignment): 0.0173 (kappa = 1 )
#> * screening p-value: 0.74257 
#> * screening method: weighted_permutation_correlation 
#> 
#> RESULT: NOT REJECTED. This is a screening result, not proof that confounding is absent.
#> NOTE: Your effect estimate must exceed the Noise Floor (delta_bound) to be meaningful.

Summary: Unified Sensitivity Analysis

The causaldef approach provides a unified framework:

┌─────────────────────────────────────────────────────────────────┐
│                     SENSITIVITY ANALYSIS                         │
├─────────────────────────────────────────────────────────────────┤
│  confounding_frontier()                                          │
│    → Maps δ as function of confounding strength (α, γ)          │
│    → Benchmarks observed covariates as reference points         │
├─────────────────────────────────────────────────────────────────┤
│  nc_diagnostic()                                                 │
│    → Empirical falsification test                               │
│    → Bounds δ using observable negative control                 │
├─────────────────────────────────────────────────────────────────┤
│  policy_regret_bound()                                           │
│    → Translates δ into decision-theoretic consequences          │
│    → Transfer penalty = Mδ; minimax floor = (M/2)δ              │
└─────────────────────────────────────────────────────────────────┘

Key Advantages:

  1. Decision-theoretic meaning: δ bounds actual regret, not just association strength
  2. Multi-method comparison: See which adjustment does best
  3. Empirical validation: Negative controls test assumptions
  4. Visual sensitivity: Frontiers show robustness at a glance

References

  1. Akdemir, D. (2026). Constraints on Causal Inference as Experiment Comparison. DOI: 10.5281/zenodo.18367347

  2. VanderWeele, T. J., & Ding, P. (2017). Sensitivity Analysis in Observational Research: Introducing the E-value. Annals of Internal Medicine.

  3. Cinelli, C., & Hazlett, C. (2020). Making Sense of Sensitivity: Extending Omitted Variable Bias. JRSS-B.

  4. Torgersen, E. (1991). Comparison of Statistical Experiments. Cambridge University Press.

These binaries (installable software) and packages are in development.
They may not be fully stable and should be used with caution. We make no claims about them.