Skip to content
D-LIT Logo

Techniques & Models

Process optimization, simulation, forecasting, and bottleneck analysis.

By D-LIT Team

Operational data without analytical technique produces dashboards full of numbers and managers unable to act on them. The techniques covered in this article are the methods that convert raw operational data (from ERP, MES, IoT systems, and process event logs) into decisions that reduce cost, increase throughput, and improve quality.

This is the deepest article in the operational analytics section. It covers manufacturing-origin techniques that have become standard practice, and extends into two areas that most frameworks leave underserved: process mining for business process analysis, and service operations analytics for non-manufacturing organizations. For the data sources that feed these techniques, see Operational Data Sources. For the KPIs these techniques support, see Operational KPIs.

OEE Decomposition Analysis

Overall Equipment Effectiveness is the starting point for most manufacturing efficiency initiatives, but its analytical value comes from decomposition rather than the top-line number.

The three-factor decomposition:

OEE = Availability x Performance x Quality

Availability = (Planned Production Time - Downtime) / Planned Production Time
Performance  = (Actual Output / Theoretical Max Output During Available Time)
Quality      = (Good Units / Total Units Started)

A facility at 65 percent OEE has the same top-line score regardless of whether the loss is driven by equipment failures (Availability), speed losses (Performance), or defects (Quality). The management actions for each are entirely different.

Availability losses decompose into planned downtime (scheduled maintenance, changeovers, breaks) and unplanned downtime (breakdowns, material shortages, operator absence). Planned downtime can be reduced through changeover optimization (SMED methodology) and maintenance scheduling. Unplanned downtime requires root cause analysis of failure events.

Performance losses decompose into speed losses (running below rated speed) and minor stops (brief stoppages under five to ten minutes that are often not recorded as downtime). Speed losses are frequently hidden: operators reduce line speed to avoid quality problems or equipment stress without formally recording a deviation. Detecting speed losses requires comparing actual output count during available time to the rated speed output count, a calculation that depends on reliable IoT or MES cycle count data.

Minor stops are the most underreported category of loss. A line that stops for 90 seconds, 40 times per shift, loses an hour of production invisibly. Automated OEE systems that capture every stop, regardless of duration, consistently find that minor stop losses account for 15 to 25 percent of total production time in high-complexity discrete manufacturing.

Quality losses decompose into startup rejects (product produced during process stabilization that does not meet specification) and steady-state rejects (defects produced during stable running conditions). These have different causes and different interventions. Startup rejects indicate a process capability problem or changeover standard problem. Steady-state rejects indicate a process drift problem or incoming material quality problem.

Practical OEE analysis workflow:

  1. Calculate OEE at the asset level for each shift and production run.
  2. Decompose into the three factors to identify the primary loss category.
  3. Within the primary loss category, rank loss events by total time impact (not by frequency; a breakdown that lasts four hours matters more than a minor stop that occurs twenty times).
  4. For the top three to five loss events by time impact, initiate root cause analysis.
  5. Implement countermeasures, standardize the improvement, and measure OEE recalculation.

This cycle (measure, decompose, prioritize, root cause, improve, confirm), executed consistently, drives OEE improvement of 10 to 20 percentage points over 12 to 18 months in most manufacturing environments.

Statistical Process Control

Statistical Process Control (SPC) is the use of statistical methods to monitor a process and detect when it moves outside acceptable variation bounds. SPC distinguishes between common cause variation (inherent to the process and unavoidable without redesign) and special cause variation (attributable to a specific, identifiable event that warrants investigation and correction).

Control charts are the core SPC tool. The most widely used types:

X-bar and R charts: Monitor the mean and range of a process characteristic measured from subgroups. The X-bar chart tracks whether the process mean is shifting. The R chart tracks whether process variability is changing. Control limits are set at ±3 standard deviations from the process mean, calculated from historical stable process data.

Upper Control Limit (UCL) = X-bar-bar + A2 x R-bar
Lower Control Limit (LCL) = X-bar-bar - A2 x R-bar

Where A2 is a control chart constant that depends on subgroup size.

Individuals and Moving Range (I-MR) charts: Used when subgrouping is not possible, for example when one unit is produced per measurement opportunity. The I chart monitors individual values; the MR chart monitors the moving range between consecutive values.

P-charts and NP-charts: Used for attribute data (pass/fail, conforming/nonconforming). The P-chart monitors the proportion defective; the NP-chart monitors the count of defectives when subgroup size is constant.

Interpreting SPC signals:

A process is out of control (special cause present) when any of the following patterns appear on a control chart:

  • One point beyond a control limit.
  • Two of three consecutive points in the outer third of the chart (beyond 2 sigma from center line).
  • Four of five consecutive points in the outer two-thirds of the chart (beyond 1 sigma from center line).
  • Eight or more consecutive points on the same side of the center line (a run).
  • Six or more consecutive points trending in one direction (a trend).

Each of these patterns signals that something has changed in the process. The appropriate response is to stop, investigate, and identify the cause before the process drifts further or produces additional defects.

Process Capability Analysis:

Where control charts monitor whether a process is stable over time, process capability analysis measures whether a stable process can meet customer specifications.

Cp  = (USL - LSL) / (6 x sigma)      [Capability, assuming centered process]
Cpk = min[(USL - mean) / (3 x sigma), (mean - LSL) / (3 x sigma)]  [Capability, accounting for centering]

A Cpk of 1.0 means the process is just barely capable of meeting specifications (0.27 percent defect rate). A Cpk of 1.33 (the common target) corresponds to a 64 ppm defect rate. A Cpk of 1.67 corresponds to 0.57 ppm. Six Sigma processes achieve a short-term Cpk of 2.0; when accounting for the standard 1.5 sigma long-term shift, the long-term Cpk is 1.5.

Lean Operations Analytics

Lean manufacturing analytics quantifies the cost and impact of the eight categories of waste (muda) that lean methodology identifies: overproduction, waiting, transportation, over-processing, inventory, motion, defects, and unused talent. The analytical objective is to measure waste in cost or time terms so that improvement projects can be prioritized by return on effort.

Value Stream Mapping with data:

Value Stream Mapping (VSM) is typically presented as a qualitative workshop exercise. With operational data, it becomes a quantitative analysis. For each step in the process:

Value-Added Time: Time during which the product is being transformed in a way the customer values.
Non-Value-Added Time: Time in queues, waiting, transport, inspection, and rework.

Process Efficiency = Value-Added Time / Total Lead Time x 100

In most discrete manufacturing operations, process efficiency (the percentage of total lead time that is actually value-adding) is 5 to 20 percent. In transactional and service processes, it is often lower. This ratio identifies the theoretical improvement ceiling from lead time reduction initiatives.

Inventory analytics:

Work-in-process (WIP) inventory is both a measure of operational health and a symptom of process problems. Little’s Law provides the fundamental relationship:

Lead Time = WIP / Throughput Rate

This means that for a given throughput rate, reducing WIP directly reduces lead time, and conversely, that high lead times are always associated with high WIP. Monitoring WIP levels by process step reveals where inventory is accumulating, which identifies bottlenecks and constraint steps.

Changeover time analysis (SMED):

Single Minute Exchange of Die (SMED) methodology separates changeover activities into internal elements (must be performed with equipment stopped) and external elements (can be performed while equipment is running). Analytical support for SMED requires time-study data or video analysis:

Changeover Time Reduction Potential = Sum of External Elements Misclassified as Internal

Converting internal elements to external elements and streamlining remaining internal elements typically reduces changeover time by 30 to 50 percent in the first pass, which directly increases OEE Availability.

Predictive Maintenance Analytics

Predictive maintenance (PdM) uses equipment sensor data (vibration, temperature, current draw, acoustic signatures, pressure) to detect early signs of equipment degradation before failure occurs. The objective is to replace time-based preventive maintenance schedules with condition-based maintenance: performing maintenance when the equipment indicates it is needed, not on a calendar schedule.

The business case:

Time-based maintenance is inherently wasteful. Equipment maintained on a fixed schedule is sometimes replaced before it has failed (premature replacement) and sometimes fails between scheduled maintenance windows (unplanned downtime). Predictive maintenance eliminates both categories of waste.

Typical outcomes from mature PdM programs: 15 to 25 percent reduction in maintenance cost, 10 to 20 percent reduction in unplanned downtime, and 2 to 5 percent improvement in OEE Availability.

Feature engineering from sensor data:

Raw sensor time series must be transformed into features that indicate equipment health. Common features:

  • Root Mean Square (RMS) of vibration signal: indicates overall vibration energy
  • Crest Factor = Peak Vibration / RMS: elevated crest factor indicates impulsive events (common in bearing defects)
  • Kurtosis: statistical measure of distribution tail weight, sensitive to early bearing defects
  • Spectral analysis: frequency-domain features that reveal characteristic defect frequencies (bearing inner race, outer race, rolling element defects each produce distinct frequency signatures)

Model types:

Threshold-based models: flag when a feature exceeds a defined limit. Simple to implement and explain, but require domain expertise to set appropriate thresholds and are not adaptive to changing operating conditions.

Anomaly detection models: learn the normal operating envelope from historical data and flag deviations. More adaptive, but can produce false positives when operating conditions change legitimately (e.g., running at a different speed or load).

Remaining Useful Life (RUL) models: predict the time remaining before failure using regression or survival analysis. The most operationally useful output (“this bearing will fail in 12 to 15 days”) but requires sufficient historical failure data to train accurately.

Implementation path:

The predictive maintenance analytics implementation sequence typically follows: establish sensor data capture and historian → validate data quality and completeness → build baseline equipment health profiles from stable operating periods → define alert thresholds or train anomaly detection models → validate against historical failure events → integrate with maintenance work order system → measure reduction in unplanned downtime and maintenance cost.

Root Cause Analysis Methods

Root cause analysis (RCA) is the structured process of identifying the fundamental cause of a defect, failure, or process excursion, as opposed to its symptom. Effective RCA prevents recurrence; ineffective RCA produces corrective actions that address symptoms and allow the root cause to produce additional events.

Fishbone (Ishikawa) Diagram:

The fishbone diagram organizes potential causes into six standard categories: Machine, Method, Material, Measurement, Man (People), and Environment (the 6Ms). It is a brainstorming and categorization tool, not a data analysis tool in itself. Its analytical value comes from pairing it with data: after generating hypotheses in the six categories, each hypothesis should be tested against operational data.

5 Whys analysis:

The 5 Whys method traces causality backward by repeatedly asking why each observed effect occurred. It is most effective for well-defined single-cause problems. For complex problems with multiple interacting causes, fault tree analysis or fishbone combined with data validation is more appropriate.

Pareto analysis:

Pareto analysis ranks defect types, failure modes, or non-conformance categories by frequency or cost impact.

Cumulative Impact % = Running Sum of Individual Category Impact / Total Impact x 100

The objective is to identify the 20 percent of causes that drive 80 percent of impact, directing improvement resources to the categories with the highest return. Pareto analysis is most useful when the categorization is specific enough to imply a cause: “machine stop” is not useful; “Conveyor 7 jam - transfer section” is actionable.

Capacity Planning Analytics

Capacity planning analytics answers the question of whether the operation can absorb projected demand, and if not, where additional capacity must be created and at what cost.

Constraint identification:

The Theory of Constraints identifies that any system has exactly one constraint, the step that limits throughput for the system as a whole. Improvement anywhere other than the constraint does not improve system throughput. The analytical task is to identify the constraint step with certainty.

The constraint is typically identified by the process step with the highest utilization, the largest WIP queue upstream, or the longest average waiting time. When multiple candidates exist, time-series analysis of WIP behavior at each step confirms which step accumulates inventory during demand surges. The accumulation point is the constraint.

Capacity modeling:

Available Capacity (units/period) = (Working Hours x Efficiency %) / Cycle Time per Unit
Required Capacity               = Demand Forecast (units/period) / Yield Rate
Capacity Gap                    = Required Capacity - Available Capacity

Capacity modeling should include efficiency assumptions (actual OEE or utilization, not theoretical) and yield assumptions (actual first pass yield, not 100 percent). Models built on theoretical capacity consistently understate the true gap and produce investment decisions that fail to close the shortage.

Scenario modeling:

Capacity planning models become most valuable when used for scenario analysis: what is the capacity impact of a 20 percent demand increase? What happens to lead time and cost per unit if a second shift is added? What is the capacity effect of a 5 percentage point improvement in OEE? Scenario modeling allows operations leaders to evaluate options before committing capital.

Process Mining

Process mining is the analytical technique of reconstructing actual business process flows from event log data in enterprise systems. It is one of the most powerful and underused techniques in operational analytics, applicable to any process that generates a digital record of activities.

Most process analytics relies on participants describing how a process works, in workshops, interviews, or process maps. Process mining bypasses this entirely by letting the data show how the process actually operates. The result is often dramatically different from the official process model.

How process mining works:

Every enterprise system (ERP, CRM, ticketing system, loan origination system, HR platform) records events in a log that captures: a case ID (which process instance), an activity (what happened), and a timestamp (when it happened). With these three elements, process mining algorithms reconstruct the actual process flow, including all the variants, detours, rework loops, and bypassed steps that the official process map does not show.

Process discovery:

Process discovery algorithms (Alpha algorithm, Heuristics Miner, Inductive Miner) take the event log and produce a process model that represents all observed behavior, including the happy path and all deviations. A process that was designed with five sequential steps often has twenty or thirty observed variants in practice, including cases that loop back to earlier steps, cases that skip mandatory steps, and cases that follow paths never documented.

Conformance checking:

Once the intended process model is known, conformance checking compares each case’s actual event sequence to the intended sequence and flags deviations. This produces a compliance rate and identifies which specific deviations are most frequent:

Conformance Rate = Cases Matching Intended Process Flow / Total Cases x 100
Deviation Impact = Deviation Frequency x Average Cost or Cycle Time Premium per Deviation

Performance analysis:

With timestamps on every event, process mining calculates the time spent in each activity and between activities for every case. This produces:

  • Average and distribution of waiting time at each process step
  • Cycle time by case variant (cases that follow the happy path versus cases that loop)
  • Bottleneck identification based on actual waiting time distributions rather than assumed processing times

Where process mining applies:

Order-to-cash: How do sales orders actually flow through the organization from entry to payment? Where do they stall, loop back, or require manual intervention? Which customers or order types generate the most deviation?

Purchase-to-pay: What is the actual flow of purchase requisitions through approval and payment? Where are invoices held? Which vendor or category combinations generate the most exceptions?

Incident management: How do IT or operational incidents actually flow from detection through resolution? Which incidents loop through multiple reassignments? Which resolution paths correlate with faster closure?

Loan or application processing: What is the actual sequence of underwriting and approval steps? Where do applications stall? Which applicant profiles or product types generate the most rework?

Process mining tooling:

Purpose-built process mining platforms (Celonis, Signavio Process Intelligence, Minit, Disco/Fluxicon) provide the discovery, conformance checking, and performance analysis capabilities with direct connectors to major enterprise systems. For organizations with existing data infrastructure, open-source libraries (PM4Py in Python) enable process mining on extracted event logs without dedicated tooling.

Service Operations Analytics

Service operations (contact centers, field service, professional services, financial processing, healthcare operations, software delivery) have their own analytical vocabulary and metric set that manufacturing-origin frameworks do not cover. This section fills that gap.

Contact center analytics:

The contact center is one of the highest-instrumented service environments, generating data on every interaction: talk time, hold time, after-call work, disposition codes, transfers, escalations, and customer satisfaction scores.

Workforce management analytics for contact centers centers on three relationships:

The Erlang C model (and its variants) describes the relationship between call arrival rate, average handle time, and required staffing to achieve a target service level:

Service Level = Probability that a call is answered within T seconds
             = f(Traffic Intensity, Number of Agents, Target Time T)

Traffic Intensity (Erlangs) = Arrival Rate x Average Handle Time

Erlang-based planning is the foundation of contact center workforce management. Over-staffing relative to Erlang requirements drives labor cost above the efficient frontier. Under-staffing drives service level below target and pushes handle time up (as agents rush calls, generating more repeat contacts).

Adherence and utilization analytics:

Schedule Adherence = Actual Available Time / Scheduled Available Time x 100
Occupancy         = Handle Time / (Handle Time + Available Idle Time) x 100

Sustainable occupancy in a contact center is typically 80 to 88 percent, depending on interaction type. Above 88 percent, queue lengths grow nonlinearly (the queueing theory effect), wait times increase, and agent burnout accelerates. Below 75 percent, labor efficiency is degraded.

First Contact Resolution analytics:

FCR is the metric with the highest correlation to both customer satisfaction and cost efficiency in contact center operations. Each repeat contact is a cost multiplier and a satisfaction detractor.

Measuring FCR requires either survey-based feedback (asking customers if their issue was resolved) or operational proxies (repeat contacts on the same issue within a defined window, typically 7 days).

FCR Rate   = Contacts Resolved on First Interaction / Total Contacts x 100
Repeat Rate = Contacts on Same Issue within N Days / Total Contacts x 100

FCR analysis by issue type, agent group, and channel identifies where resolution capability is strongest and where process or knowledge gaps are causing repeat contacts.

Field service analytics:

Field service organizations (utilities, HVAC, medical equipment, telecommunications, property management) manage a mobile workforce whose time is split between travel, on-site work, and administrative tasks. Field service analytics tracks:

Utilization Rate = Billable or Productive Hours / Total Available Hours x 100
First-Time Fix Rate = Jobs Completed Without Return Visit / Total Jobs x 100
Mean Time to Repair (MTTR) = Total Repair Time / Number of Repair Events
Travel Time % = Travel Time / Total Time x 100

First-time fix rate is the field service equivalent of First Pass Yield. A low first-time fix rate indicates either a parts availability problem (technician arrives without the correct part), a skills problem (technician cannot diagnose or repair on the first visit), or a diagnostic problem (dispatch sends the wrong technician or provides insufficient pre-visit information).

Scheduling optimization analytics for field service uses the operational data (job durations, travel time matrices, technician skill sets, SLA requirements) to determine optimal daily route and job assignment, a vehicle routing problem with time windows that is tractable at operational scale with modern optimization solvers.

Professional services analytics:

Professional services organizations (consulting, legal, accounting, engineering) are fundamentally resource-allocation businesses. Their operational analytics centers on utilization, margin by engagement, and delivery performance.

Billable Utilization = Billable Hours / Total Available Hours x 100
Realization Rate     = Revenue Billed / Revenue at Standard Rate x 100
Margin per Engagement = (Revenue - (Direct Labor Cost + Allocated Overhead)) / Revenue x 100

Project-level analytics tracks planned versus actual hours by phase, scope change frequency, and milestone adherence. Operations leaders use this data to identify which project types, which client segments, and which delivery models generate the most predictable margin and which consume disproportionate overhead.

Software delivery operations (DevOps) analytics:

Software engineering organizations have developed their own operational metric framework. The DORA (DevOps Research and Assessment) metrics measure delivery performance:

Deployment Frequency   = Number of Production Deployments / Time Period
Lead Time for Changes  = Commit Time to Production Deployment Time
Change Failure Rate    = Failed Deployments / Total Deployments x 100
Mean Time to Recovery  = Total Recovery Time / Number of Incidents

These four metrics are the operational KPIs of software delivery. High-performing engineering organizations achieve multiple deployments per day, lead times measured in hours rather than weeks, change failure rates below 15 percent, and recovery times under one hour. The analytics infrastructure to measure and trend these metrics is an operational analytics program for engineering the same way OEE analytics is an operational analytics program for manufacturing.

Integrating Techniques into an Operational Analytics Practice

No single technique covers the full scope of operational analysis. A mature operational analytics program uses techniques in combination, matched to the decision type:

  • Real-time monitoring uses SPC control charts and OEE dashboards to detect variance as it occurs.
  • Shift-level analysis uses lean analytics and bottleneck analysis to understand where production is behind plan and why.
  • Weekly reviews use Pareto analysis and root cause investigation to prioritize the top-impact problems.
  • Monthly planning uses capacity modeling and scenario analysis to assess resource requirements against the demand forecast.
  • Continuous improvement programs use process mining to identify systemic process deviations that accumulate into significant cost and cycle time impact.
  • Asset management programs use predictive maintenance analytics to optimize maintenance spend and reduce unplanned downtime.

The selection and integration of these techniques, supported by the right data sources and communicated through the right dashboards, constitutes a fully operational analytics capability. For the dashboard architectures that surface these insights to the people who need them, see Dashboards and Reporting.

Get More from D-LIT

Ready to transform your analytics capabilities? Talk to our team about how D-LIT can help your organisation make better, data-driven decisions.

Get in Touch