

This visitor put up is written by Dr. Julian Runge, an Assistant Professor in Built-in Advertising and marketing Communications at Northwestern College, and William Grosso, the CEO of Recreation Knowledge Professionals.
Observational Causal Inference (OCI) seeks to determine causal relationships from observational information, when no experimental variation and randomization are current. OCI is utilized in digital product and advertising analytics to infer the impression of various methods on outcomes like gross sales, buyer engagement, and product adoption. OCI generally fashions the connection between variables noticed in real-world information.
In advertising, probably the most widespread functions of OCI is in Media and Advertising and marketing Combine Modeling (m/MMM). m/MMM leverages historic gross sales and advertising information to estimate the impact of assorted actions throughout the advertising combine, equivalent to TV, digital advertisements, promotions, pricing, or product adjustments, on enterprise outcomes. Hypothetically, m/MMM allows firms to allocate budgets, optimize campaigns, and predict future advertising and product efficiency. m/MMM usually makes use of regression-based fashions to estimate these impacts, assuming that different related elements are both managed for or could be accounted for by means of statistical strategies.
Nonetheless, MMM and comparable observational approaches usually fall into the entice of correlating inputs and outputs with out guaranteeing that the connection is really causal. As an illustration, if promoting spend spikes throughout a selected vacation season and gross sales additionally rise, an MMM may attribute this improve to promoting, even when it was primarily pushed by seasonality or different exterior elements.
Observational Causal Inference Usually Fails to Determine True Results
Regardless of its widespread use, a rising physique of proof signifies that OCI methods usually stray from appropriately figuring out true causal results. This can be a vital problem as a result of incorrect inferences can result in misguided enterprise selections, leading to monetary losses, inefficient advertising methods, or misaligned product improvement efforts.
Gordon et al. (2019) present a complete critique of promoting measurement fashions in digital promoting. They spotlight that the majority OCI fashions are susceptible to endogeneity (the place causality flows in each instructions between variables) and omitted variable bias (the place lacking variables distort the estimated impact of a therapy). These points aren’t simply theoretical: the research finds that fashions incessantly misattribute causality, resulting in incorrect conclusions in regards to the effectiveness of promoting interventions, highlighting a have to run experiments as an alternative.
A newer research by Gordon, Moakler, and Zettelmeyer (2023) goes a step additional, demonstrating that even subtle causal inference strategies usually fail to duplicate true therapy results when in comparison with outcomes from randomized managed trials. Their findings name into query the validity of many generally used enterprise analytics methods. These strategies, regardless of their complexity, usually yield biased estimates when the assumptions underpinning them (e.g., no unobserved confounders) are violated—a standard prevalence in enterprise settings.
Past the context of digital promoting, a current working paper by Bray, Sanders and Stamatopoulos (2024) notes that “observational worth variation […] can not reproduce experimental worth elasticities.” To contextualize the severity of this downside, think about the context of medical trials in medication.


When a brand new drug is examined, RCTs are the gold customary as a result of they get rid of bias and confounding, making certain that any noticed impact is really attributable to the therapy. Nobody would belief observational information alone to conclude {that a} new remedy is secure and efficient. So why ought to companies belief OCI methods when tens of millions of {dollars} are at stake in digital advertising or product design?
Certainly, OCI approaches in enterprise usually depend on assumptions which can be simply violated. As an illustration, when modeling the impact of a worth change on gross sales, an analyst should assume that no unobserved elements are influencing each the value and gross sales concurrently. If a competitor launches the same product throughout a promotion interval, failing to account for this may possible result in overestimating the promotion’s effectiveness. Such flawed insights can immediate entrepreneurs to double down on a method that’s ineffective and even detrimental in actuality.
Prescriptive Suggestions from Observational Causal Inference Could Be Misinformed
If OCI methods fail to determine therapy results appropriately, the scenario could also be even worse on the subject of the insurance policies these fashions inform and advocate. Enterprise and advertising analytics aren’t simply descriptive—they usually are used prescriptively. Managers use them to resolve the right way to allocate tens of millions in advert spend, the right way to design and when to run promotions, or the right way to personalize product experiences for customers. When these selections are primarily based on flawed causal inferences, the enterprise penalties might be extreme.
A main instance of this problem is in m/MMM, the place advertising measurement not solely estimates previous efficiency however additionally instantly informs an organization’s actions for the subsequent interval. Suppose an m/MMM incorrectly estimates that rising spend on show advertisements drives gross sales considerably. The agency could resolve to shift extra price range to show advertisements, doubtlessly diverting funds from channels like search or TV, which can even have a stronger (however underestimated) causal impression. Over time, such misguided actions can result in suboptimal advertising efficiency, deteriorating return on funding, and distorted assessments of channel effectiveness. What’s extra, because the fashions fail to precisely inform enterprise technique, government confidence in m/MMM methods could be considerably eroded.
One other context the place flawed OCI insights can backfire is in personalised UX design for digital merchandise like apps, video games, and social media. Firms usually use data-driven fashions to find out what kind of content material or options to current to customers, aiming to maximise engagement, retention, or conversion. If these fashions incorrectly infer {that a} sure characteristic causes customers to remain longer, the corporate may overinvest in enhancing that characteristic whereas neglecting others which have a real impression. Worse, they might even make adjustments that cut back person satisfaction and drive churn.
The Drawback Is Severe – And Its Extent At the moment Not Absolutely Appreciated
Nascent large-scale real-world proof means that, even when OCI is carried out on huge, wealthy, and granular datasets, the core problem of incorrect estimates stays. Opposite to common perception, having extra information doesn’t clear up the elemental problems with confounding and bias. Gordon et al. (2023) present that rising the quantity of knowledge with out experimental validation doesn’t essentially enhance the accuracy of OCI methods. It could even amplify biases, making analysts extra assured in flawed outcomes.
The important thing level to restate is that this: With out experimental validation, OCI is susceptible to being incorrect, both in magnitude or in signal. That’s, the mannequin could not simply fail to measure the dimensions of the impact appropriately—it could even get the course of the impact mistaken. An organization may find yourself reducing a channel that’s really extremely worthwhile or investing closely in a method that has a destructive impression. In the end, that is the worst-case situation for an organization deeply embracing data-driven decision-making.
Mitigation Methods
Given the restrictions and dangers related to OCI, what can firms do to make sure they make selections knowledgeable by sound causal insights? There are a number of remedial methods.
Probably the most easy answer is to conduct experiments wherever attainable. A/B assessments, geo-based experiments, and incrementality assessments can all assist set up causality with excessive confidence. (For a choice tree guiding your selection of methodology, please see Determine 1 right here.)


For digital merchandise, RCTs are sometimes possible: for instance, testing completely different variations of an online web page or various the concentrating on standards for advertisements. Working experiments, even on a small scale, can present floor reality for causal results, which might then be used to validate or calibrate observational fashions.
One other strategy are bandit algorithms that conduct randomized trials together with coverage studying and execution. Their means to study insurance policies “on the go” is the important thing benefit they create. This nonetheless requires lots of premeditation and cautious planning to leverage them efficiently. We need to point out them right here, however advise to begin with easier approaches to get began with experimentation.
In actuality, operating experiments (or bandits) throughout all enterprise areas isn’t all the time sensible or attainable. To assist be certain that OCI fashions produce correct estimates for these conditions, you possibly can calibrate observational fashions utilizing experimental outcomes. For instance, if a agency has run an A/B take a look at to measure the impact of a reduction marketing campaign, the outcomes can be utilized to validate an m/MMM’s estimates of the identical marketing campaign. This course of, referred to as calibrating observational fashions with experimental benchmarks, helps to regulate for biases within the observational estimates. This text in Harvard Enterprise Evaluation summarizes other ways how calibration could be carried out, emphasizing the necessity for steady validation of observational fashions utilizing RCTs. This iterative course of ensures that the fashions stay grounded in correct empirical proof.
In sure situations, you might be extremely assured that the assumptions for OCI to provide legitimate causal estimates are met. An instance might be the outcomes of a tried-and-tested attribution mannequin. Calibration and validation of OCI fashions in opposition to such outcomes will also be a wise technique.
One other associated strategy could be to develop a devoted mannequin that’s skilled on all obtainable experimental outcomes to supply causal assessments throughout different enterprise analytics selections and use circumstances. In a means, such a mannequin could be framed as a “causal attribution mannequin.”
In some conditions, experiments and calibrations might not be possible as a result of price range constraints, time limitations, or operational challenges. In such circumstances, we advocate utilizing well-established enterprise methods to cross-check and validate coverage suggestions derived from OCI. If the fashions’ inferences aren’t aligned with these methods, double- and triple-check. Examples for such methods are:
- Pricing: Buy historical past, geo-location, or value-based pricing fashions which were extensively validated within the tutorial literature
- Promoting Methods: Concentrate on good inventive methods that align together with your model values somewhat than blindly following mannequin outputs
- Product Growth: Prioritize options and functionalities primarily based on confirmed theories of shopper habits somewhat than purely data-driven inferences
By leaning into time-tested methods, companies can reduce the danger of adopting flawed insurance policies urged by doubtlessly biased fashions.
If unsure, err on the facet of warning and follow a presently profitable technique somewhat than implementing ineffective or dangerous adjustments. For current computational advances on this regard, check out the m/MMM package deal Robyn. It supplies the means to formalize a desire for non-extreme outcomes along with experiment calibration in a multi-objective optimization framework.
A Name to Motion: Experiment, Calibrate, Validate
In conclusion, whereas OCI methods are beneficial for exploratory evaluation and producing hypotheses, present proof means that counting on them with out additional validation is dangerous. In advertising and enterprise analytics, the place selections instantly impression income, model fairness, and buyer experiences, companies can not afford to behave on deceptive insights.
“Combating Misinformation” could also be a powerful body for our name to motion. Nonetheless, even misinformation on social media is typically shared with out the originator understanding the knowledge is fake. Equally, a knowledge scientist who invested weeks of labor into OCI-based modeling could deeply imagine within the accuracy of their outcomes. These outcomes would nonetheless nonetheless misinform enterprise selections with potential to negatively impression share- and stakeholders.
To keep away from pricey errors, firms ought to deal with OCI as a place to begin, not the ultimate phrase.


Wherever attainable, run experiments to validate your fashions and calibrate your estimates. If experimentation isn’t possible, be vital of your fashions’ outputs and all the time cross-check with established enterprise methods and inside experience. With out such safeguards, what you are promoting technique might be constructed on misinformation, resulting in misguided selections and wasted sources.
And what higher time to problem this name, with the Convention on Digital Experimentation (CODE) at MIT taking place later this week. CODE gathers each the utilized and tutorial analytics neighborhood to dive deep into experimentation as a pillar of enterprise and advertising analytics. We hope to see you there.
About Julian and Invoice
Julian Runge is a behavioral economist and information scientist. He’s presently an Assistant Professor of Advertising and marketing at Northwestern College. Beforehand, Julian labored as a researcher on recreation information science and advertising analytics at Northeastern, Duke and Stanford College, and at Fb. Julian has printed extensively on these subjects within the proceedings of premier machine studying conferences equivalent to IEEE COG and AAAI AIIDE, and in main journals equivalent to Data Methods Analysis, Quantitative Advertising and marketing and Economics and Harvard Enterprise Evaluation.
William Grosso is an entrepreneur and investor primarily based in San Mateo, California. Over his profession, Grosso has labored for quite a lot of expertise firms and is the founding father of a number of startups, together with Scientific Income, which pioneered dynamic pricing in cellular video games, and Recreation Knowledge Professionals which focuses on income optimization in digital leisure. Grosso is thought for his experience in distributed methods, income optimization, and information science, and has given talks on these subjects at conferences world wide. He holds a grasp’s diploma in arithmetic from UC Berkeley and has labored as a analysis scientist in Synthetic Intelligence at Stanford College. He’s the creator or co-author of three books on software program improvement and over 50 scientific papers.
Photographs by Michał Parzuchowski, Jason Dent, and Nathan Dumlao on Unsplash