Unraveling the Power of Causal Machine Learning
Understanding Causality
Standard machine learning is very good at finding patterns. Given enough data, a model can identify that A and B tend to co-occur, that X is predictive of Y, that a certain feature combination correlates with higher churn. What it cannot tell you, by itself, is why — whether A causes B, or both are driven by some third variable C.
This matters the moment you want to make a decision. If you’re considering an intervention — running a marketing campaign, changing a dosage, implementing a policy — correlation is often not enough. You need to know what will happen when you act, not just what tends to be true in the historical data.
Causal inference tries to answer those “what if” questions: what would the outcome have been had we treated differently? What’s the effect of this intervention, holding everything else constant?
The Challenge of Causal Inference
The core difficulty is the fundamental problem of causal inference: you can never observe the same unit under both treatment and control at the same time. You either give the patient the drug or you don’t. You can’t observe both counterfactuals for the same person simultaneously.
Observational data makes this harder because the people who receive a treatment are typically different from those who don’t — in ways that also affect the outcome. This confounding is the reason we can’t simply compare treated and untreated groups and call the difference the causal effect.
Causal Machine Learning
Causal Machine Learning sits at the intersection of causal inference and predictive modeling. The goal is to use flexible machine learning models to estimate causal effects, while respecting the assumptions needed for those estimates to be valid.
Methods
Randomized Control Trials (RCTs) remain the cleanest approach. By randomly assigning units to treatment or control, you break the link between treatment assignment and confounders. Any difference in outcomes can then be attributed to the treatment. The limitation is practical: RCTs are expensive, slow, and sometimes impossible or unethical to run.
Propensity Score Matching (PSM) is used when you have observational data and can’t randomize. The propensity score is the estimated probability of receiving treatment given observed covariates. By matching treated and control units with similar propensity scores, you create a quasi-experimental comparison that controls for observed confounders. The key word is observed — PSM can’t control for confounders you haven’t measured.
Instrumental Variables (IV) handle a trickier case: you have unmeasured confounders, but you can find a variable (the instrument) that affects the treatment but has no direct effect on the outcome. Using the instrument as a source of quasi-random variation in treatment, you can isolate the causal effect. Finding credible instruments is often the hard part.
Synthetic Control constructs a counterfactual for the treated unit using a weighted combination of control units. It’s particularly useful for case studies — a single country or region that adopted a policy, where you want to estimate what would have happened without it. The synthetic control is designed to match the treated unit’s pre-treatment trajectory as closely as possible, making the post-treatment divergence attributable to the intervention.
Applications
Causal ML has found real applications in public health, economics, and marketing. In healthcare, it supports personalized treatment recommendations — estimating not just the average treatment effect but the individual-level effect, which can vary dramatically across patients. In marketing, it helps separate the effect of a campaign from the background tendency of customers to convert anyway.
Reinforcement learning is a natural extension of this thinking: instead of estimating the effect of a single intervention, you’re learning a policy for sequential decision-making under uncertainty. That’s a topic for a future post.