Skip to content
Concepts & Methodology

What Is Predictive Quality in Mfg Test

Predictive quality uses production data to catch defects before they happen. Learn how it works, what data it needs, and how test results feed prediction.

JJulien Buteau
intermediate8 min readMarch 14, 2026

What Is Predictive Quality in Manufacturing Test

Predictive quality uses historical test data and process data to forecast defects before they occur. Instead of catching bad units at end-of-line testing, predictive quality identifies the conditions that produce bad units and flags them earlier in the process. This guide covers how predictive quality works, what data it needs, and how structured test results feed prediction models.

Reactive vs Predictive Quality

ApproachWhen Defects Are FoundCost
Reactive (inspect and reject)After the unit is builtHigh (scrap, rework, field returns)
Statistical (SPC)When a process drifts out of controlMedium (catches trends, not individual units)
Predictive (ML-based)Before the defect occursLow (prevents defective units from being built)

Traditional quality control catches defects after they happen. SPC catches trends before they produce defects. Predictive quality goes further: it uses patterns in upstream data to predict which specific units or batches will fail downstream tests.

How Predictive Quality Works

The core idea: upstream test data contains signals that correlate with downstream failures.

StepWhat Happens
1. Collect dataEvery measurement, at every test stage, for every unit
2. Find correlationsML models identify which upstream measurements predict downstream failures
3. Build prediction modelTrain a model on historical pass/fail data with upstream features
4. Deploy inlineRun predictions on live test data as units move through production
5. Act on predictionsFlag high-risk units for additional inspection or route to rework

Example

A power supply fails output ripple testing at end-of-line. Analysis of historical data shows that units with input capacitor ESR above 45 milliohms at incoming inspection are 8x more likely to fail ripple testing. A predictive model flags these units at IQC before they're assembled.

What Data Feeds Predictive Quality

Data SourceWhat It ProvidesStage
Incoming quality (IQC)Component measurements, supplier lot dataBefore assembly
In-process quality (IPQC)SPI paste volume, AOI defect counts, reflow profileDuring assembly
Functional test (FCT)Electrical measurements with limitsAfter assembly
End-of-line test (EOL)Final pass/fail, measurement valuesBefore shipping
Environmental dataTemperature, humidity on the production floorContinuous
Equipment dataFixture cycle count, instrument calibration ageContinuous

The more structured your test data, the better the predictions. Measurements with units, limits, and serial number traceability are the foundation.

Predictive Quality Use Cases

Use CaseInput DataPredictionBenefit
Skip testingUpstream measurementsUnit will pass downstream test10-50% test time reduction
Early warningProcess trendsBatch will have high failure rateCatch before full batch is built
Supplier qualityIQC measurements by vendorLot will cause downstream failuresReject lots at incoming
Limit optimizationMeasurement distributionsCurrent limits are too wide or too tightReduce false failures and escapes
Field failure predictionProduction test dataUnit will fail within warranty periodTighten limits or add screening

Prerequisites

  • Python 3.10+
  • OpenHTF installed (pip install openhtf)
  • TofuPilot Python SDK installed (pip install tofupilot)

Step 1: Capture Structured Test Data

Predictive quality starts with clean, consistent test data. Every measurement needs a name, value, unit, and limits.

production_test.py
import openhtf as htf
from openhtf.util import units


@htf.measures(
    htf.Measurement("input_capacitor_esr_mOhm")
    .in_range(maximum=50)
    .with_units(units.OHM),
    htf.Measurement("output_ripple_mV")
    .in_range(maximum=30)
    .with_units(units.MILLIVOLT),
    htf.Measurement("efficiency_percent")
    .in_range(minimum=90)
    .with_units(units.PERCENT),
)
def phase_electrical_test(test):
    """Capture measurements that feed predictive models."""
    test.measurements.input_capacitor_esr_mOhm = 38.2
    test.measurements.output_ripple_mV = 22.1
    test.measurements.efficiency_percent = 93.4

Step 2: Log Everything to TofuPilot

Every run uploads to TofuPilot with serial number, measurements, limits, and pass/fail status. This structured data is what predictive models need.

production_test.py
from tofupilot.openhtf import TofuPilot

test = htf.Test(phase_electrical_test)

with TofuPilot(test):
    test.execute(test_start=lambda: input("Scan serial: "))

Step 3: Use TofuPilot Data for Analysis

TofuPilot tracks measurement distributions, correlations, and trends across all test stages. Open the Analytics tab to identify:

  • Measurement correlations between upstream and downstream tests
  • Failure patterns by supplier lot, station, or time period
  • Distribution shifts that precede failure rate increases
  • Marginal results that predict future failures

This data is the starting point for building predictive models. The structured format (measurements with units, limits, and serial traceability) eliminates the data cleaning step that typically consumes 80% of ML project time.

Predictive Quality vs Traditional Quality

AspectTraditionalPredictive
When defects are foundAfter they happenBefore they happen
Test strategyTest everything the same wayAdapt testing based on risk
Data useCompliance reportingReal-time decision making
Limit settingFrom datasheet or engineering judgmentFrom production data and ML models
ROI measurementDefect rate reductionCost avoidance (prevented scrap, skipped tests)

Getting Started

Predictive quality doesn't require a massive ML infrastructure on day one. The progression is:

LevelWhat You DoWhat You Need
1. CollectLog all measurements with units and limitsOpenHTF + TofuPilot
2. VisualizeReview distributions, correlations, and trendsTofuPilot Analytics
3. CorrelateIdentify upstream predictors of downstream failuresExport data, run correlation analysis
4. PredictBuild and deploy ML models on live dataData science team + production data
5. AutomateAct on predictions inline (skip tests, flag units)Model deployment infrastructure

Most teams get significant value from levels 1-3 alone. Just visualizing the correlation between incoming component measurements and end-of-line failures reveals actionable insights without building any ML models.

More Guides

Put this guide into practice