What Is Predictive Quality in Manufacturing Test
Predictive quality uses historical test data and process data to forecast defects before they occur. Instead of catching bad units at end-of-line testing, predictive quality identifies the conditions that produce bad units and flags them earlier in the process. This guide covers how predictive quality works, what data it needs, and how structured test results feed prediction models.
Reactive vs Predictive Quality
| Approach | When Defects Are Found | Cost |
|---|---|---|
| Reactive (inspect and reject) | After the unit is built | High (scrap, rework, field returns) |
| Statistical (SPC) | When a process drifts out of control | Medium (catches trends, not individual units) |
| Predictive (ML-based) | Before the defect occurs | Low (prevents defective units from being built) |
Traditional quality control catches defects after they happen. SPC catches trends before they produce defects. Predictive quality goes further: it uses patterns in upstream data to predict which specific units or batches will fail downstream tests.
How Predictive Quality Works
The core idea: upstream test data contains signals that correlate with downstream failures.
| Step | What Happens |
|---|---|
| 1. Collect data | Every measurement, at every test stage, for every unit |
| 2. Find correlations | ML models identify which upstream measurements predict downstream failures |
| 3. Build prediction model | Train a model on historical pass/fail data with upstream features |
| 4. Deploy inline | Run predictions on live test data as units move through production |
| 5. Act on predictions | Flag high-risk units for additional inspection or route to rework |
Example
A power supply fails output ripple testing at end-of-line. Analysis of historical data shows that units with input capacitor ESR above 45 milliohms at incoming inspection are 8x more likely to fail ripple testing. A predictive model flags these units at IQC before they're assembled.
What Data Feeds Predictive Quality
| Data Source | What It Provides | Stage |
|---|---|---|
| Incoming quality (IQC) | Component measurements, supplier lot data | Before assembly |
| In-process quality (IPQC) | SPI paste volume, AOI defect counts, reflow profile | During assembly |
| Functional test (FCT) | Electrical measurements with limits | After assembly |
| End-of-line test (EOL) | Final pass/fail, measurement values | Before shipping |
| Environmental data | Temperature, humidity on the production floor | Continuous |
| Equipment data | Fixture cycle count, instrument calibration age | Continuous |
The more structured your test data, the better the predictions. Measurements with units, limits, and serial number traceability are the foundation.
Predictive Quality Use Cases
| Use Case | Input Data | Prediction | Benefit |
|---|---|---|---|
| Skip testing | Upstream measurements | Unit will pass downstream test | 10-50% test time reduction |
| Early warning | Process trends | Batch will have high failure rate | Catch before full batch is built |
| Supplier quality | IQC measurements by vendor | Lot will cause downstream failures | Reject lots at incoming |
| Limit optimization | Measurement distributions | Current limits are too wide or too tight | Reduce false failures and escapes |
| Field failure prediction | Production test data | Unit will fail within warranty period | Tighten limits or add screening |
Prerequisites
- Python 3.10+
- OpenHTF installed (
pip install openhtf) - TofuPilot Python SDK installed (
pip install tofupilot)
Step 1: Capture Structured Test Data
Predictive quality starts with clean, consistent test data. Every measurement needs a name, value, unit, and limits.
import openhtf as htf
from openhtf.util import units
@htf.measures(
htf.Measurement("input_capacitor_esr_mOhm")
.in_range(maximum=50)
.with_units(units.OHM),
htf.Measurement("output_ripple_mV")
.in_range(maximum=30)
.with_units(units.MILLIVOLT),
htf.Measurement("efficiency_percent")
.in_range(minimum=90)
.with_units(units.PERCENT),
)
def phase_electrical_test(test):
"""Capture measurements that feed predictive models."""
test.measurements.input_capacitor_esr_mOhm = 38.2
test.measurements.output_ripple_mV = 22.1
test.measurements.efficiency_percent = 93.4Step 2: Log Everything to TofuPilot
Every run uploads to TofuPilot with serial number, measurements, limits, and pass/fail status. This structured data is what predictive models need.
from tofupilot.openhtf import TofuPilot
test = htf.Test(phase_electrical_test)
with TofuPilot(test):
test.execute(test_start=lambda: input("Scan serial: "))Step 3: Use TofuPilot Data for Analysis
TofuPilot tracks measurement distributions, correlations, and trends across all test stages. Open the Analytics tab to identify:
- Measurement correlations between upstream and downstream tests
- Failure patterns by supplier lot, station, or time period
- Distribution shifts that precede failure rate increases
- Marginal results that predict future failures
This data is the starting point for building predictive models. The structured format (measurements with units, limits, and serial traceability) eliminates the data cleaning step that typically consumes 80% of ML project time.
Predictive Quality vs Traditional Quality
| Aspect | Traditional | Predictive |
|---|---|---|
| When defects are found | After they happen | Before they happen |
| Test strategy | Test everything the same way | Adapt testing based on risk |
| Data use | Compliance reporting | Real-time decision making |
| Limit setting | From datasheet or engineering judgment | From production data and ML models |
| ROI measurement | Defect rate reduction | Cost avoidance (prevented scrap, skipped tests) |
Getting Started
Predictive quality doesn't require a massive ML infrastructure on day one. The progression is:
| Level | What You Do | What You Need |
|---|---|---|
| 1. Collect | Log all measurements with units and limits | OpenHTF + TofuPilot |
| 2. Visualize | Review distributions, correlations, and trends | TofuPilot Analytics |
| 3. Correlate | Identify upstream predictors of downstream failures | Export data, run correlation analysis |
| 4. Predict | Build and deploy ML models on live data | Data science team + production data |
| 5. Automate | Act on predictions inline (skip tests, flag units) | Model deployment infrastructure |
Most teams get significant value from levels 1-3 alone. Just visualizing the correlation between incoming component measurements and end-of-line failures reveals actionable insights without building any ML models.