MES Software: Vendors, Features & Costs Compared 2026
MES software compared: vendors, functions per VDI 5600, costs (cloud vs. on-premise) and implementation. Honest market overview 2026.
Production metrics are the quantified measurements a manufacturing organisation uses to understand, control and improve its operations. They translate the messy reality of the shopfloor — machines starting and stopping, parts being made, defects being produced, operators changing over tooling, energy being consumed — into a small set of numbers that can be compared, trended and acted upon. Done well, production metrics are the instrument panel of a factory. Done badly, they are the manufacturing equivalent of a speedometer that reads whatever the driver wants to believe.
After 25 years of manufacturing work across four continents, including three years running DMAIC projects as a Six Sigma Black Belt in automotive headliner production and a decade leading global MES and traceability programmes at Johnson Controls and Visteon, I have formed one strong conviction about production metrics. It is the same conviction I wrote up as a narrative in my book OEE: One Number, Many Lies: most factories do not have a KPI problem, they have a KPI honesty problem. The numbers reported to management are almost always better than the numbers reality would produce. The difference between the two is where all the improvement potential lives.
This article is the long-form English counterpart to the German blog article on the same topic. It covers what production metrics actually are, the canonical KPI set every discrete manufacturing operation should know, the frameworks (ISO 22400, VDI 3423, SQDCP) that give these numbers structure, the honest benchmarks you will not find in vendor brochures, and the specific ways production numbers get quietly distorted — so you can recognise the patterns in your own data.
Not all production metrics serve the same purpose. One of the more common causes of KPI confusion in manufacturing is reporting the same number to three different audiences who need three different things from it. A useful KPI system separates its metrics into three levels, each with its own update frequency, audience and purpose.
| Level | Audience | Update frequency | Example metrics |
|---|---|---|---|
| Strategic | Plant manager, COO, executive board | Monthly / quarterly | Plant OEE, cost per unit, on-time delivery, revenue per employee, CO₂ per unit |
| Tactical | Production manager, shift supervisors, CI leads | Daily / weekly | Line OEE, FPY, scrap rate by part, MTBF, changeover time, schedule adherence |
| Operational | Operators, team leaders, maintenance | Real-time / per cycle | Current cycle time vs. target, live downtime reason, parts vs. plan, rejects in last hour |
The common failure mode is forcing one level of metric to serve another audience. A plant manager who is asked to react to real-time line stops drowns in noise; an operator who receives only monthly OEE has no way to improve in the moment. The best-run factories I have seen build their KPI system from the operational layer up: if operators cannot see and influence the numbers in real time, nothing the strategic layer reports will actually change.
There is no universal KPI list that fits every factory. But there is a canonical set of about 15 metrics that, together, cover 90 % of what any discrete manufacturing operation needs. The ISO 22400 standard — Automation systems and integration — Key performance indicators for manufacturing operations management — formalises 34 KPIs, but in practice most plants operate well on the subset below.
| Category | Metric | What it actually measures | Typical formula |
|---|---|---|---|
| Availability | Availability rate | Share of planned production time actually running | Run time / Planned production time |
| MTBF | Mean time between failures | Total run time / number of failures | |
| MTTR | Mean time to repair / restore | Total downtime / number of failures | |
| Changeover time | Non-productive time between two orders | Last-good-part to first-good-part | |
| Performance | Performance rate | Speed loss vs. ideal cycle time | (Ideal cycle × total count) / Run time |
| Cycle time (actual) | Average time per part produced | Run time / total parts | |
| Throughput | Parts produced per unit of time | Good parts / shift (or hour) | |
| Quality | Quality rate | Share of parts conforming first time | Good parts / total parts |
| Scrap rate | Share of irrecoverable defects | Scrap parts / total parts | |
| FPY / RFT | First-pass yield (right first time) | Parts good without rework / total | |
| Composite | OEE | Combined availability × performance × quality | A × P × Q |
| Flow & delivery | On-time delivery (OTD) | Orders shipped within commitment | On-time orders / total orders |
| Schedule adherence | Execution vs. plan per shift/day | Executed / planned orders | |
| WIP / Takt adherence | Work-in-process vs. target; takt deviation | WIP count; actual cycle vs. takt | |
| Resource | Energy per unit | kWh or MJ per part (line or plant) | Energy consumed / good parts |
| Labour productivity | Output per direct labour hour | Good parts / direct labour hours |
Most factories can manage operational excellence with this set. Adding metric #16, #17 and #18 rarely produces proportional insight; it usually produces dashboard bloat. One of the most underrated KPI skills is knowing when to stop adding metrics.
Every production metric is either a lagging indicator (describing what already happened) or a leading indicator (predicting what will happen). Most KPI systems overweight the lagging side. This is intuitive — lagging metrics are easier to measure and harder to argue with — but it systematically limits the organisation's ability to intervene in time.
| Lagging indicator | Corresponding leading indicator |
|---|---|
| Monthly OEE | Live micro-stops per hour, SPC out-of-control signals |
| Customer complaints (PPM) | Internal scrap rate, Cpk/Ppk trend, MSA status |
| On-time delivery | Schedule adherence, WIP vs. target, changeover time |
| Warranty cost | FPY, SPC trend on critical characteristics |
| Unplanned downtime cost | Vibration/current/temperature anomalies (condition data) |
The single most effective shift I have seen plants make in 25 years is the move from monthly OEE reports to live OEE visibility with structured downtime capture. The OEE number itself does not improve because of this change; but the leading indicators it exposes — micro-stops, reason-coded stops, speed deviations — become actionable. The number follows.
Every MES vendor can produce a slide where "customers achieve OEE above 85 %." This is true in carefully selected cases. The honest distribution across real manufacturing plants is different. The numbers below are aggregated from direct experience across 900+ machines globally at Johnson Controls and Visteon, cross-checked against public industry studies, and from what SYMESTIC observes across 15,000+ connected machines.
| KPI | Typical starting point | Good industry level | World-class |
|---|---|---|---|
| OEE (discrete manufacturing) | 45–60 % | 65–75 % | ≥ 85 % |
| Availability | 70–80 % | 85–90 % | > 92 % |
| Performance | 75–85 % | 90–95 % | > 95 % |
| Quality / FPY | 95–98 % | 98–99,5 % | > 99,7 % |
| Scrap rate (automotive) | 2–5 % | 0,5–1,5 % | < 0,3 % |
| Customer PPM (automotive) | > 200 | 25–100 | < 10 |
| On-time delivery | 75–85 % | 92–97 % | > 98 % |
| Schedule adherence | 60–75 % | 85–92 % | > 95 % |
An observation that recurs in almost every plant I have worked in: the OEE number drops by 15–20 percentage points in the first 4–8 weeks after honest, automated measurement is introduced. The plant did not get worse. It got measured correctly for the first time. The previous "80 % OEE" was a combination of selective manual data capture, micro-stops not being counted, and unrecorded short speed losses. This first drop is predictable and should be built into the change-management narrative from the start, or the metric will be declared a failure exactly at the moment it is finally working.
Three frameworks appear repeatedly in any serious discussion of production metrics. Each solves a different problem.
| Framework | What it is | Why it matters |
|---|---|---|
| ISO 22400 | International standard defining 34 manufacturing KPIs with precise formulas | Ends the "what exactly is OEE" debate; enables cross-plant and cross-company comparability |
| VDI 3423 | German guideline on availability and time-model definitions for production equipment | Standardises how planned, unplanned and organisational time is treated — the root of most OEE disputes |
| SQDCP | Shopfloor-management KPI pillars: Safety, Quality, Delivery, Cost, People | Organises the dashboard around what actually needs daily management attention, not by technical hierarchy |
My strong recommendation for any plant serious about production metrics: adopt ISO 22400 definitions for the core KPIs and SQDCP as the dashboard structure. The two complement each other — ISO defines how to calculate, SQDCP defines how to present. What they share is the refusal to let KPIs drift in definition from shift to shift, which is the single most common source of metric-based disputes between production and management.
This is the section most KPI articles avoid. It is also the most important. The real problem with production metrics is not usually the formula; it is the data upstream of the formula. Here is the honest catalogue of distortions I have seen repeatedly across four continents.
| Distortion pattern | How it works | How to detect it |
|---|---|---|
| Ideal-cycle-time creep | Nominal cycle time quietly adjusted upward to make performance look better | Compare current ideal cycle to engineering specification or PPAP document |
| Downtime reclassification | Unplanned stops booked as planned, changeover, or "organisational" time | Sudden shifts in downtime category distribution month-over-month |
| Micro-stop invisibility | Stops shorter than X minutes excluded from capture | Compare count-based throughput to time-based calculation — the gap reveals micro-stops |
| Scrap hidden in rework | Defective parts rebooked through repair loops instead of scrap | Rework-to-first-pass ratio trend; rework labour hours vs. plan |
| Planned-time gaming | Availability calculated against a shortened "planned production time" | Plot planned production time as a KPI itself over time |
| Order-end timing | Last parts of a good batch run against next order to "pre-bank" output | Compare order completion timestamps to physical shift transitions |
| Definition drift | "OEE" silently redefined across shifts, plants, or years | Audit the actual formula in use against ISO 22400 or documented company standard |
None of this is usually malicious. It is the natural consequence of measuring people with numbers they can partially influence, while giving them the ability to influence the numbers themselves. The defence against it is not tighter management — it is taking the measurement out of human hands altogether. Automated data capture from PLCs, sensors and MES solves the distortion problem at the source. This is the single strongest structural argument for a modern MES: it is not about new insights, it is about honest numbers.
One pattern recurs so often in my work that I now treat it as a rule: the quality of a production metric is determined by the quality of the data that feeds it, not the sophistication of the formula. A plant with automated PLC-based capture of downtime, parts produced, and cycle times, using the simplest OEE formula, will outperform a plant with a perfectly engineered KPI hierarchy running on paper-based data. The formula is the easy part. The data is the hard part.
This is why I always recommend the same sequence to plants starting a production-metrics programme:
Inverting this sequence — setting targets before establishing clean measurement — is the most common and most expensive mistake in KPI programmes. It creates exactly the distortion patterns listed above, because people are asked to hit numbers that the measurement system cannot truthfully produce.
| Role | Core questions the dashboard must answer | Key metrics |
|---|---|---|
| Operator | Am I hitting takt? What is my current order status? Where did the last stop come from? | Live cycle time vs. target, parts vs. plan, last stop reason, current OEE |
| Shift supervisor | Which lines are on plan? Where do I need to intervene in the next 2 hours? | Line-level OEE, top 3 downtime reasons today, quality alerts, schedule adherence |
| Production manager | Are we on track this week? What are the recurring loss drivers? | Plant-level OEE trend, Pareto of losses, FPY trend, schedule adherence |
| Plant manager / COO | Are we delivering the plan? Is the trend improving? Where is the structural risk? | SQDCP scorecard, cost per unit, OTD, CO₂ per unit, labour productivity |
The most common dashboard mistake — and I have seen it in nine out of ten plants I have assessed — is showing every role the same KPI set, usually the executive scorecard. Operators receive monthly OEE and cannot act on it. Plant managers receive live cycle times and drown in noise. The measurement infrastructure needs to be unified; the presentation layer must not be.
Some metrics are tracked in most plants because they have always been tracked, not because they drive decisions. The honest delete list:
Replacing vanity metrics with action metrics — not "how many 5-Why sessions did we run" but "what % of recurring defects were root-caused within the target time" — is often the single highest-leverage change in a mature KPI system.
| Dimension | Without MES | With SYMESTIC MES |
|---|---|---|
| Data source | Paper, Excel, manual transcription | Automatic capture from PLC, OPC UA, digital I/O |
| Update frequency | Shift-end, day-end, month-end | Real-time, per cycle |
| Definition consistency | Drifts across shifts/plants | Configured centrally, applied everywhere identically |
| Human distortion | Structural — metrics reflect the reporter as much as reality | Eliminated at source; system captures before humans see |
| Root-cause analysis | Reconstructed from memory and paper logs | Timestamp-joined data across machine state, order, material, operator |
The Meleghy case is a concrete example: after SYMESTIC went live at six plants across four countries, the automotive supplier achieved a 10 % reduction in downtime, 7 % improvement in output and 5 % improvement in availability — not because the plants suddenly got better at making parts, but because for the first time, everyone saw the same honest numbers at the same time. The Klocke case is a faster version of the same story: 12 % output improvement and 8 % availability improvement within three weeks, on pharma packaging lines connected through digital I/O gateways with no LAN infrastructure. In both cases, the KPIs did not change. The measurement changed.
How many production metrics should a factory track?
Fewer than most factories do. For operational management, 8–12 well-defined and well-measured KPIs outperform 30+ partially measured ones. A useful test: can every KPI on your dashboard be linked to a specific person accountable for it, and a specific action it would trigger if it moved? If not, it is information, not a KPI. The canonical set of OEE with its three components, FPY, scrap rate, MTBF, MTTR, changeover time, schedule adherence, on-time delivery, and energy per unit covers 90 % of what discrete manufacturing actually needs to manage. Everything beyond that is either industry-specific (batch tracking in pharma, SPC characteristics in precision machining) or strategic reporting. The temptation to add "just one more metric" is constant; resisting it is one of the quieter disciplines of a mature operations organisation.
What is the right frequency for reporting production metrics?
Different for each layer. Operators need real-time or near-real-time — seeing what happened five minutes ago does not help them fix what is happening now. Shift supervisors need shift-boundary summaries plus live alerts on exceptions. Production managers need daily roll-ups with weekly trend context. Plant managers need weekly performance against plan with monthly trend analysis. The trap almost every plant falls into is running one reporting cadence — usually weekly or monthly — and making everyone consume it. Operators lose engagement because the data is historical; executives get overwhelmed because the data is too granular. A modern MES with live dashboards and scheduled roll-ups solves this by letting the same underlying data appear in the right cadence at every level automatically.
Which is more important: leading or lagging indicators?
Both, in a deliberate ratio. A well-designed KPI system carries roughly 60–70 % leading indicators at the operational and tactical layers — the metrics operations can actually influence in time — and 30–40 % lagging indicators at the strategic layer to confirm that the leading metrics are driving the outcomes they should. Most plants invert this ratio, reporting 80 %+ lagging indicators because they are easier to measure and harder to argue about. The result is a KPI system that is accurate but not actionable: everyone knows what went wrong yesterday; no one can change what will go wrong tomorrow. Moving even 20 percentage points of the KPI mix from lagging to leading — from "last month's OEE" to "this hour's micro-stop count" — typically delivers more improvement than adding any new metric.
Why do production metrics almost always get worse when measurement is improved?
Because the old numbers were wrong. The unpleasant but consistent finding across hundreds of plants: reported performance exceeds real performance by 10–20 percentage points in most operations before honest measurement is introduced. This is not conscious fraud; it is the accumulated effect of unrecorded micro-stops, reclassified downtime, optimistically set ideal cycle times, and the natural tendency to round up when filling in paper forms at shift end. When an MES replaces this with automatic capture, the first 4–8 weeks look like performance collapsed. It did not; it finally became visible. Plants that understand this pattern in advance — and build it into the change-management narrative — transition through it smoothly. Plants that do not often cancel the measurement project exactly at the moment it started working, because the new numbers are politically unacceptable. The honest number is always worth more than the comfortable one; but the organisation needs to agree on that truth before the data arrives, not after.
How does SYMESTIC approach production metrics differently?
By treating the measurement layer as the product, not the dashboards. Most competitors invest heavily in visualisation and surprisingly little in ensuring that the underlying data is complete, honest and reason-coded. SYMESTIC inverts this: the platform is built to capture parts, cycle times, stops, alarms and process data automatically through OPC UA, MQTT or digital I/O gateways — including on machines from the 1990s that were never designed for it — so the numbers that reach the dashboard are the numbers reality produced, not the numbers someone decided to enter. On top of this foundation sit the standard KPIs (OEE per ISO 22400, FPY, MTBF, changeover, schedule adherence, energy per unit), the SQDCP-aligned dashboards for each role, and the integration points that push data into SAP, ERP and CMMS systems. The operational result, consistent across 15,000+ connected machines and 18 countries: productivity improvement measurable within the first 12 weeks, and an organisation that for the first time operates from one shared, honest picture of its own production. The point of the metric is never the metric; it is the decision the metric enables. Get the measurement honest, and the decisions improve on their own.
Related: OEE · Quality Control · Machine Downtime · Predictive Maintenance · Six Sigma · Kaizen · Production Costs · Production Metrics · MES
MES software compared: vendors, functions per VDI 5600, costs (cloud vs. on-premise) and implementation. Honest market overview 2026.
OEE software captures availability, performance & quality automatically in real time. Vendor comparison, costs & case studies. 30-day free trial.
MES (Manufacturing Execution System): Functions per VDI 5600, architectures, costs and real-world results. With implementation data from 15,000+ machines.