Skip to content
D-LIT Logo

Techniques & Models

Segmentation, churn prediction, propensity modeling, and sentiment analysis.

By D-LIT Team

The analytical techniques in this guide are the methods that transform customer data from a reporting resource into a competitive advantage. Applied well, they let you predict which customers will leave before they do, identify which customers will expand and when, and focus customer success, product, and marketing resources on the actions most likely to change outcomes.

The techniques range from foundational (cohort analysis and RFM segmentation can be executed in a well-structured SQL environment without machine learning) to advanced, where probabilistic lifetime value models and churn propensity scoring require statistical or ML infrastructure. This guide covers both ends of the spectrum, explains the underlying mechanics in terms relevant to executive decision-making, and addresses the significant methodological differences between B2B and B2C contexts that most practitioners overlook.


Cohort Analysis

Cohort analysis is the foundational technique of customer retention measurement. It groups customers by when they were acquired (the acquisition month or quarter is the most common grouping) and tracks what proportion of each cohort remains active at each subsequent time interval. The resulting cohort retention table or retention curve reveals whether your retention is improving, degrading, or stable over time, and whether apparent overall retention health is masking deterioration in newer cohorts.

Why aggregate retention rates mislead. An aggregate monthly retention rate of 90 percent can be consistent with dramatically different underlying realities. In a growing business, newer cohorts with worse retention are diluted by the weight of older, longer-tenured customers. The aggregate looks stable while the product experience may be deteriorating for every new customer you acquire. Cohort analysis surfaces this by tracking each cohort independently.

How to build a cohort retention table. The standard output is a matrix where rows represent acquisition cohorts (e.g., “Customers who first activated in January 2024”) and columns represent time intervals since acquisition (Month 0, Month 1, Month 2, and so on). Each cell contains the percentage of the original cohort still active at that time interval.

Cohort     | Month 0 | Month 1 | Month 3 | Month 6 | Month 12
Jan 2024   | 100%    | 78%     | 64%     | 55%     | 44%
Feb 2024   | 100%    | 81%     | 68%     | 59%     | 48%
Mar 2024   | 100%    | 84%     | 71%     | 63%     | 52%

The improvement from January to March cohorts suggests that product or onboarding changes made during that period improved early retention. This is the kind of signal that an aggregate retention rate cannot produce.

Revenue cohort analysis. The same technique applied to revenue rather than customer count produces Gross Revenue Retention (GRR) and Net Revenue Retention (NRR) cohort curves. NRR above 100 percent in a cohort means that the cohort’s revenue has grown over time from expansion, more than offsetting any churned revenue within the cohort. NRR is one of the most powerful indicators of a business with strong product-market fit and effective expansion motion.

What to do with the findings. Cohort analysis findings have direct implications for product roadmap prioritization, onboarding redesign, and customer success resource allocation. If Month 1 retention is the dominant drop-off point, the investment case for improving onboarding is clear and quantifiable. If Month 6 is the critical drop-off, the case is for proactive mid-lifecycle customer success intervention.


RFM Segmentation

RFM segmentation classifies customers along three dimensions: Recency (how recently they purchased or engaged), Frequency (how often they purchase or engage), and Monetary value (how much revenue they generate). The combination of these three scores produces a customer prioritization matrix that commercial and customer success teams can act on directly.

How RFM scoring works. Each customer receives a score on each dimension, typically on a scale of 1 to 5, based on their position in the distribution relative to the full customer base. A customer who purchased recently, purchases frequently, and generates high revenue receives a score of 5-5-5. A customer who last purchased long ago, purchases rarely, and generates low revenue receives a score of 1-1-1.

RFM Score = Recency Score (1-5) + Frequency Score (1-5) + Monetary Score (1-5)

Or alternatively, treated as a three-digit composite: a customer scoring 5-5-5 is labeled a “Champions” segment; one scoring 1-1-1 is labeled “Lost.”

Actionable RFM segments. The value of RFM is not the scoring itself but the segment definitions that emerge and the differentiated treatment each segment warrants.

  • Champions (High R, High F, High M): Your most valuable customers. They should receive the lightest-touch customer success engagement and the highest priority for product beta programs and advisory boards.
  • Loyal Customers (High F, High M, Moderate R): Consistent buyers who have not purchased very recently. Re-engagement campaigns and loyalty programs are appropriate.
  • At-Risk Customers (High F historically, but low recent Recency): Previously valuable customers whose engagement is dropping. Proactive outreach with personalized messaging referencing their specific usage history is warranted.
  • Cannot Lose Them (High M, Low R): High-value customers who have become inactive. These accounts justify direct executive-level engagement.
  • Lost (Low R, Low F, Low M): Inactive low-value customers where the cost of re-engagement likely exceeds the expected return.

B2B adaptation. In B2B SaaS, RFM must be adapted because “purchase frequency” does not have the same meaning as in transaction-based businesses. Useful B2B substitutes: Recency maps to last product login or last meaningful feature use; Frequency maps to number of active users on the account or number of product sessions per week; Monetary maps to contract value or MRR. The same segmentation logic applies; the inputs reflect the subscription rather than transaction model.


Churn Prediction and Propensity Modeling

Churn prediction is the analytical technique with the highest direct revenue impact in most subscription businesses. A model that accurately identifies at-risk customers 60 to 90 days before their renewal date gives customer success teams the lead time to intervene effectively. Without prediction, customer success is reactive, responding to customers who have already decided to leave.

The two approaches: rule-based and statistical.

Rule-based churn scoring assigns risk scores based on predefined thresholds applied to customer health signals. Examples: an account that has not logged in for 14 days receives 20 risk points; an account that has submitted two or more support tickets in the past 30 days receives 15 risk points; a declining trend in feature adoption over the past 60 days receives 25 risk points. The total score produces a risk tier. This approach is transparent, auditable, and can be implemented without data science infrastructure. Its limitation is that the thresholds and weights are assumptions, not learned from data.

Statistical and machine learning approaches learn the relationship between behavioral signals and churn outcomes from historical data. Logistic regression is the standard starting point: it produces a probability that a given customer will churn in the next N days, based on a set of input features. More complex approaches including gradient boosting models (XGBoost, LightGBM) and neural networks can capture nonlinear relationships and interactions between features that logistic regression misses.

Feature engineering for churn models. The predictive power of a churn model depends heavily on the features provided to it. The most predictive features typically fall into these categories:

  • Engagement trajectory: Not just current engagement level, but the trend over the past 30, 60, and 90 days. A customer whose logins have declined 40 percent over three months is more at risk than one who has always had low engagement.
  • Feature adoption breadth: Customers using only one or two features are more vulnerable than those with deep product adoption across multiple modules.
  • Support history: Volume and recency of support tickets, particularly those tagged as unresolved or escalated.
  • Contract signals: Time to renewal, contract tier, whether the customer has expanded or contracted their contract in the past.
  • Onboarding completion: Whether the customer completed key activation milestones in their first 30 days.
  • NPS and survey history: Detractor classification or recent decline in CSAT is a strong predictor of churn.
  • Stakeholder engagement: In B2B accounts, whether the primary champion is still active in the product or has left the company.

Model validation. A churn model must be validated against held-out historical data before being deployed in production. Key metrics: the model’s precision (of customers it flags as at-risk, what proportion actually churned?) and recall (of customers who actually churned, what proportion did it flag?). A model with high precision but low recall misses many churners; one with high recall but low precision generates excessive false positives that consume customer success capacity without return. The AUC-ROC score provides a single summary metric for discriminative performance.

Operationalizing churn predictions. A churn model that produces scores but does not trigger customer success actions has no business value. The operational integration is as important as the model quality. Standard integration patterns include: automated alerts in Salesforce or HubSpot when an account crosses a risk threshold, weekly churn risk review queues assigned to customer success managers, and account health scoring dashboards that surface model predictions in the customer success team’s workflow.

The intervention problem. Churn prediction enables intervention, but the intervention must be effective to generate return. Common intervention approaches include: proactive executive business reviews for high-value at-risk accounts, targeted feature adoption campaigns for accounts underusing core modules, loyalty and pricing offers for price-sensitive accounts approaching renewal, and escalation to technical resources for accounts with unresolved product issues. Measuring intervention effectiveness, by comparing churn rates of intervened accounts against a comparable control group, is essential to understanding which interventions actually work.


Lifetime Value Modeling

Lifetime value modeling quantifies the expected net revenue that a customer relationship will generate over its duration. This figure is foundational to acquisition economics, retention investment decisions, and customer segmentation strategy.

The simple model. The widely used formula for LTV in subscription businesses is:

LTV = (Average Revenue per Account per Month x Gross Margin %) / Monthly Churn Rate

This model assumes a constant churn rate and constant revenue, which are simplifications. It is useful as a first approximation but understates the value of customers who expand over time and overstates the value of customers whose churn probability varies across the lifecycle.

Accounting for expansion revenue. A more accurate model uses Net Revenue Retention rather than gross churn:

Adjusted LTV = (ARPA x Gross Margin %) / (Churn Rate - Expansion Rate)

If monthly churn is 2 percent but monthly expansion from existing customers is 0.5 percent, the effective net retention rate is 98.5 percent and the denominator is 1.5 percent rather than 2 percent. This significantly increases the calculated LTV and is more accurate for businesses with strong upsell motion.

Important limitation: This formula is only valid when churn rate exceeds expansion rate. When expansion equals or exceeds churn, meaning Net Revenue Retention is at or above 100 percent, the denominator becomes zero or negative, producing infinite or negative LTV, which is nonsensical. Businesses with NRR above 100 percent should use a discounted cash flow approach over a fixed time horizon instead:

Bounded LTV = Sum over t=1 to N of: (ARPA x Gross Margin % x (1 + Expansion Rate - Churn Rate)^t) / (1 + Discount Rate)^t

This caps the projection at a realistic horizon (typically 5 to 7 years for SaaS businesses) and applies a discount rate to account for risk and the time value of money; 20 to 25 percent annually is common for pre-scale companies. The bounded approach allows expansion to compound revenue growth within each period without producing an unbounded result, and it yields a more conservative and realistic LTV estimate even when NRR is below 100 percent.

Cohort-based LTV. Rather than using a single blended churn rate, cohort-based LTV modeling calculates LTV separately for each acquisition cohort or customer segment. This reveals that certain customer profiles, acquired through specific channels, on specific product tiers, or in specific verticals, generate dramatically different lifetime value. That segmentation should directly inform where acquisition investment is concentrated.

Probabilistic LTV modeling for transaction-based businesses. For businesses where customers make repeated discrete purchases rather than paying a recurring subscription, the Pareto/NBD (Negative Binomial Distribution) model and its successor BG/NBD (Beta Geometric/NBD) are the statistical standards. These models estimate the probability that a customer is still “alive” (not permanently lapsed) and their expected purchase frequency, given observed transaction history. Paired with a Gamma-Gamma model for monetary value, they produce individual-level CLV predictions.

The mechanics: the BG/NBD model fits two distributions: one for purchase frequency while active (Negative Binomial) and one for the probability of becoming inactive after any transaction (Beta Geometric). The model is fit on historical transaction data and produces a probability distribution over expected future transactions for each customer. This is significantly more accurate than simple average-based approaches and can be implemented using open-source libraries including the Python lifetimes package.

Using LTV for strategic decisions. LTV by acquisition channel informs marketing budget allocation. LTV by product tier informs pricing and packaging strategy. LTV by vertical informs sales territory and GTM focus. The ratio of LTV to CAC by segment is the most powerful filter for deciding where to concentrate growth investment.


Customer Journey Mapping

Customer journey mapping in an analytics context means tracing the actual behavioral paths customers take through the product and the customer relationship, not the intended path, but the measured one.

Funnel analysis. The most common implementation is funnel analysis: defining a sequence of steps that represent the intended path to a key outcome (activation, expansion, renewal) and measuring what proportion of customers complete each step in sequence and where drop-off occurs. Funnel analysis in product analytics platforms like Amplitude or Mixpanel allows teams to define funnels and measure conversion at each step, with the ability to segment by cohort, account type, or behavioral attributes.

Path analysis. More granular than funnel analysis, path analysis examines the actual sequences of actions customers take, rather than forcing a predefined sequence. This surfaces unexpected paths to value: customers who follow non-obvious routes to activation, or behaviors that precede churn that the predefined funnel does not capture.

Time-to-value analysis. A specific application of journey mapping is measuring time-to-value: the elapsed time between signup or contract start and the moment the customer completes their first meaningful value-generating action. Shortening time-to-value is one of the highest-return investments available to product and customer success teams, because it directly predicts early-stage retention.


NPS Driver Analysis

NPS driver analysis goes beyond the score to understand what factors explain why customers rate your product or company as they do. The most common approach is text analysis of verbatim responses combined with quantitative modeling.

Driver analysis with structured data. If survey responses include demographic and behavioral attributes (account size, product tier, usage level, support history), regression analysis can identify which factors are statistically associated with Promoter versus Detractor classification. This moves NPS from a trailing sentiment measure to a diagnostic tool: “Accounts that have submitted two or more unresolved support tickets in the past 90 days are 3.2x more likely to be Detractors.”

Text analysis of verbatim responses. Topic modeling and sentiment analysis applied to open-ended NPS responses can identify the themes that appear most frequently among Promoters and Detractors. At scale, this approach surfaces patterns that manual review would miss. Common implementations use LDA (Latent Dirichlet Allocation) for topic extraction or simpler keyword frequency analysis grouped by NPS category.

Closing the loop. The business value of NPS driver analysis is realized through the closed-loop process: identifying Detractors, understanding their specific concerns, routing their feedback to the responsible team, taking action, and following up. Measuring the re-survey NPS of previously Detractor accounts after intervention closes the loop quantitatively.


B2B vs. B2C: Critical Methodological Differences

Most customer analytics content treats B2B and B2C as equivalent contexts with minor surface differences. In practice, the methodological differences are significant enough that applying B2C approaches uncritically in a B2B context produces misleading results.

Unit of analysis. In B2C, the unit of analysis is the individual customer. In B2B, the unit of analysis is the account, but accounts contain multiple users, multiple stakeholders, and often multiple contracts. A B2B company might have 100 accounts, each with 10 to 200 users. Churn at the account level (failure to renew) is distinct from user-level disengagement. NPS collected from individual users must be aggregated thoughtfully to represent the account’s overall health.

Churn definition. In B2C subscription businesses, churn is typically a discrete, self-service event: the customer cancels or stops paying. In B2B, “churn” at renewal may occur after a months-long commercial negotiation. Partially churned accounts, customers who renew but at a significantly reduced contract value, are common and should be tracked separately as contraction. True churn at renewal is often preceded by months of warning signals that B2B-specific models must be designed to capture.

Relationship complexity. B2B customer relationships involve multiple stakeholders with different roles and influence. The power user who loves the product may have no influence over the renewal decision; the economic buyer who controls the budget may rarely touch the product. B2B health scoring should attempt to capture both product engagement signals and stakeholder engagement signals (executive sponsor involvement, business review attendance, reference willingness).

Sales cycle vs. usage cycle. In B2B, the commercial relationship begins before product usage: the sales cycle produces a contract, and then the product usage cycle begins. This means that cohort analysis in B2B should often use contract start date rather than first product login as the cohort anchor, and that early-stage health signals in the first 30 to 90 days of the contract are particularly predictive of renewal outcomes.

Data sparsity. B2B companies typically have fewer customers than B2C companies, which limits the statistical power of models. A B2C e-commerce company may have millions of customers, enabling sophisticated ML models with strong validation. A B2B SaaS company with 500 accounts may not have enough historical churn events to train a reliable predictive model. In this context, rule-based health scoring and domain-expert judgment are often more practical than black-box ML.


Predictive Analytics for Customer Success

The frontier of customer analytics connects predictive models to operational workflows in customer success, enabling proactive rather than reactive account management.

Account health scoring. An account health score synthesizes multiple signals (product engagement, support history, NPS, contract signals, stakeholder activity) into a single composite score. When implemented well, the health score reflects what a knowledgeable CSM would assess from reviewing all account data manually, but at a scale and consistency that humans cannot achieve.

Propensity to expand. Complementary to churn prediction, expansion propensity models identify accounts most likely to purchase additional seats, upgrade to a higher tier, or adopt new product modules. These models use similar input features to churn models (product breadth, feature adoption, engagement depth) but with a different target variable. An account that has deeply adopted one module and is growing in user count is a candidate for cross-sell; an account that has reached the limits of its current tier is a candidate for upgrade.

Predictive QBR prioritization. Quarterly business reviews consume significant customer success capacity. Predictive analytics can optimize where that capacity is deployed: ranking accounts by a combination of risk score, expansion potential, and strategic importance so that QBR scheduling reflects data-driven prioritization rather than relationship habits.

Early warning systems. An early warning system monitors a set of behavioral triggers and fires alerts when thresholds are crossed. Examples: login frequency has declined more than 30 percent over 30 days; an account’s primary champion has not logged in for 21 days; support ticket volume has doubled month-over-month. These triggers route to CSM queues with account context attached, enabling prompt and informed intervention.


For the data sources required to implement these techniques, see the Customer Data Sources guide. For the KPIs these techniques improve, see the Customer KPIs guide. For how to present the outputs of these analyses to executive and operational audiences, see the Dashboards guide.

Get More from D-LIT

Ready to transform your analytics capabilities? Talk to our team about how D-LIT can help your organisation make better, data-driven decisions.

Get in Touch