Most marketing teams running propensity models are optimizing for the wrong thing. They build sophisticated machine learning pipelines, tune hyperparameters with care, and deploy models that predict clicks, opens, or page visits with impressive accuracy. Then they wonder why campaign ROI barely moves.

The problem isn't the modeling technique. It's the outcome definition. Predicting who will click an email is not the same as predicting who will become a profitable customer. And predicting who will convert is not the same as predicting who will convert because of your marketing. These distinctions sound subtle, but they represent the difference between analytics that consume budget and analytics that generate revenue.

Building propensity models that actually drive marketing value requires getting three things right: choosing prediction targets that align with business outcomes, modeling incrementality rather than mere correlation, and deploying scores in ways that change real decisions. Each step is where most implementations quietly fail.

Outcome Selection: You Optimize What You Measure, So Measure What Matters

The most common mistake in marketing propensity modeling is treating the prediction target as an afterthought. Teams default to whatever outcome is easiest to capture in their data warehouse—email opens, ad clicks, site visits—and build models around those signals. The logic feels reasonable: engagement predicts conversion, so optimizing for engagement should improve results. But this reasoning breaks down faster than most teams realize.

Click propensity and purchase propensity are often weakly correlated. The people most likely to click your emails may be habitual browsers who never buy. The people most likely to convert after seeing an ad may have been planning to purchase anyway. When you optimize campaigns around engagement proxies, you end up targeting the wrong audiences and misallocating spend. Hal Varian has noted that one of the most valuable skills in data analysis is knowing which variable actually matters—and in marketing, that variable is almost never a click.

Choosing the right outcome means working backward from the business question. If the goal is customer acquisition, model first-purchase probability with a meaningful time window. If the goal is retention, model churn risk against a clearly defined inactivity threshold. If the goal is revenue growth, model expected customer lifetime value rather than single-transaction likelihood. Each choice reshapes the feature engineering, the training data, and ultimately which customers the model prioritizes.

One practical test: ask your stakeholders what action they would take differently based on the model's output. If a propensity score doesn't change a targeting decision, a budget allocation, or a creative strategy, the outcome you've chosen isn't connected to a real business lever. The model might be technically excellent and strategically useless. Outcome selection is a business decision dressed up as a technical one, and it deserves the same scrutiny you'd give any strategic choice.

Takeaway

The prediction target you choose determines whether your model drives revenue or just reports activity. Always work backward from the business decision the score will inform, not forward from the data that's easiest to collect.

Incrementality Modeling: Predicting Who You Can Actually Influence

Standard propensity models answer the question who is likely to convert? But marketing teams don't need to know who will convert—they need to know who will convert because of marketing intervention. The difference is incrementality, and ignoring it is how campaigns end up spending the most money on people who would have bought anyway.

Consider four customer segments in any campaign. Sure things will convert whether you market to them or not. Lost causes won't convert regardless. Sleeping dogs actually convert less when marketed to—they find the outreach annoying and disengage. And persuadables convert only when they receive the marketing touch. A traditional propensity model scores sure things highest because they have the strongest conversion signal in historical data. An incrementality model identifies persuadables—the only group where marketing spend generates actual return.

Building incrementality models requires either randomized holdout experiments or quasi-experimental methods. The gold standard is an uplift model trained on data from campaigns with proper control groups. You model the difference in conversion probability between treated and untreated populations, not just the conversion probability itself. Techniques like two-model approaches, transformed outcome methods, or causal forests each have trade-offs in complexity and data requirements. The critical prerequisite is disciplined experimentation—without control groups, you have no ground truth for incremental impact.

The business implications are significant. Organizations that shift from standard propensity to incrementality modeling routinely find that 30 to 50 percent of their high-propensity targets are sure things consuming budget without generating incremental revenue. Reallocating that spend toward persuadable segments often improves campaign ROI by double-digit percentages without increasing total marketing investment. The model doesn't have to be more accurate in a traditional sense—it just has to answer a more useful question.

Takeaway

Predicting who will convert rewards you for finding people who were going to buy anyway. Predicting who you can influence rewards you for finding people whose behavior your marketing actually changes. The second question is harder to answer but far more valuable.

Model Deployment: Scores Are Worthless Until They Change a Decision

A propensity model sitting in a Jupyter notebook is a science project. A propensity model integrated into campaign execution is a business asset. The gap between the two is where most marketing analytics initiatives stall. Deployment isn't a technical afterthought—it's the step that converts analytical insight into revenue, and it demands as much design thinking as the model itself.

Practical deployment means connecting model outputs to the systems where marketing decisions happen. That could be a marketing automation platform selecting audience segments, a bid management system adjusting spend allocation, or a personalization engine choosing which offer to present. The integration pattern matters: batch scoring works for email campaigns planned days ahead, but real-time scoring is necessary for web personalization or programmatic ad buying. Each pattern has different latency, infrastructure, and monitoring requirements.

Equally important is translating model scores into actionable decision rules. A raw probability isn't useful to a campaign manager. What's needed are clear thresholds, segment definitions, or ranked lists that plug directly into existing workflows. Should the top decile get a premium offer? Should scores below a threshold be excluded entirely? Should budget be allocated proportionally to incremental lift estimates? These rules should be designed collaboratively with marketing operations, not handed down from a data science team in isolation.

Finally, deployed models need monitoring and feedback loops. Customer behavior shifts, market conditions change, and model performance degrades over time—a phenomenon called concept drift. Tracking prediction accuracy against actual outcomes on a rolling basis lets teams detect degradation before it erodes campaign performance. The most effective organizations treat model deployment as the beginning of a cycle, not the end of a project. Scoring, measuring, learning, and retraining become a continuous process embedded in how marketing operates.

Takeaway

A model's value is zero until it changes a real decision in a real system. Design deployment around the workflow it needs to influence, not around the data science team that built it.

Propensity modeling in marketing is a solved problem technically but a persistent challenge strategically. The algorithms work. The bottleneck is asking them the right questions and connecting their answers to real operational decisions.

Start with outcome selection—ensure you're predicting something that maps to business value, not just data convenience. Move toward incrementality to distinguish influence from observation. And invest in deployment that embeds scores into the workflows where spend gets allocated and audiences get selected.

The competitive advantage isn't in having a model. It's in having a model that changes what your organization actually does.