A statistical study that quantifies the amount of measurement variation attributable to the measurement system itself, separating it into repeatability (equipment variation) and reproducibility (operator variation) components.
A Gage R&R study evaluates how much of the total observed variation in a measurement process is due to the measurement system versus the actual part-to-part variation. The study uses a designed experiment where multiple operators measure multiple parts multiple times. The resulting data is analyzed using ANOVA (Analysis of Variance) or the range method to decompose the total variation into three components: equipment variation (repeatability), operator variation (reproducibility), and part-to-part variation.
The key metric from a Gage R&R study is %GRR — the percentage of total variation or tolerance consumed by the measurement system. AIAG (Automotive Industry Action Group) guidelines classify results as: less than 10% GRR is generally acceptable, 10-30% is marginal and may be acceptable depending on the application, and greater than 30% indicates the measurement system needs improvement. The number of distinct categories (ndc) should be 5 or more for the measurement system to adequately distinguish between parts.
For calibration management, Gage R&R studies are an essential tool for validating measurement systems before they are put into production use. They are required by IATF 16949 (automotive quality), widely used in Six Sigma programs, and recommended by ISO 10012 for measurement management systems. A Gage R&R study can reveal that a measurement problem is due to operators (need training) rather than equipment (need better instruments), preventing unnecessary equipment purchases. Regular Gage R&R studies after calibration confirm that the measurement system continues to perform acceptably.
In aerospace calibration labs, Gage R&R studies are critical for torque wrench calibration systems where multiple technicians perform measurements on the same reference torque transducers. A typical study involves three operators measuring ten torque standards across their full range using the same Norbar Professional torque tester, collecting 30 data points per operator. Poor repeatability might indicate worn drive mechanisms, while high reproducibility variation could reveal inconsistent operator technique in applying torque. Medical device manufacturers conducting Gage R&R on dimensional measurement systems face stringent requirements - for instance, when validating CMM performance for measuring critical implant dimensions. A study might involve three metrologists measuring the same set of titanium hip joint components using a Zeiss Contura CMM, with measurements repeated over multiple days. If the study shows >30% of total variation attributed to the measurement system, the process capability becomes questionable. Common audit findings include incomplete operator training documentation, inadequate sampling plans, or failure to include all sources of variation (environmental conditions, fixture variations). In one case, a defense contractor failed FDA inspection because their Gage R&R study for pressure transducer calibration didn't account for different calibration technicians, masking significant reproducibility issues that affected product acceptance decisions.
ISO/IEC 17025:2017 addresses Gage R&R indirectly through Section 7.7.1 on measurement uncertainty evaluation, requiring labs to identify all sources of measurement variation. AS9100D specifically references measurement system analysis in Section 7.1.5.2, mandating statistical studies including Gage R&R for critical measurements. ISO 13485:2016 Section 7.6 requires medical device manufacturers to validate measurement processes, with Gage R&R being the preferred method. ANSI/NCSL Z540.3-2006 Section 4.3 explicitly requires measurement system variation studies for calibration processes. The GUM (ISO/IEC Guide 98-3) Type A uncertainty evaluation methodology aligns with repeatability components of Gage R&R studies. IATF 16949 Section 7.1.5.1.1 mandates measurement system analysis studies with specific acceptance criteria. During audits, assessors verify that studies include appropriate sample sizes (typically minimum 10 parts, 3 operators, 2-3 trials), proper statistical analysis methods, documented acceptance criteria (commonly <10% excellent, <30% acceptable), and evidence that unsatisfactory results triggered corrective actions. Auditors also examine whether environmental conditions, different equipment setups, and all relevant operators were included in studies.
CalibrationOS incorporates Gage R&R functionality within its Measurement System Analysis (MSA) module, automatically calculating repeatability, reproducibility, and total Gage R&R percentages according to AIAG MSA-4 methodology. The system captures operator identifications, measurement sequences, and environmental conditions during data collection, ensuring complete traceability. Built-in statistical engines compute variance components, generate control charts, and produce comprehensive reports showing %Tolerance, %Total Variation, and number of distinct categories. The software automatically flags studies exceeding user-defined acceptance criteria (typically 30%) and triggers corrective action workflows. Integration with the calibration scheduling module enables automatic MSA study reminders based on equipment criticality and usage patterns. During audits, the system generates standardized Gage R&R certificates showing statistical summaries, ANOVA tables, and graphical analysis including range charts and average charts. CalibrationOS also maintains historical MSA trending data, allowing labs to demonstrate measurement system stability over time and supporting ISO/IEC 17025 continual improvement requirements through automated drift detection algorithms.
A Gage R&R study is a statistical method that measures how much variation in a measurement process comes from the measurement system (repeatability and reproducibility) versus actual part-to-part differences.
Per AIAG guidelines, %GRR below 10% is generally acceptable, 10-30% is marginal, and above 30% indicates the measurement system needs improvement. The number of distinct categories should be 5 or more.
Analyze whether the problem is repeatability (equipment) or reproducibility (operators). Equipment issues may need better instruments or maintenance. Operator issues often require training, standardized procedures, or fixtures to reduce technique variation.
This article is licensed CC BY-SA 4.0. Share, adapt, and reuse with attribution to calibrationos.com/glossary/gage-rr.
Industry benchmarks, best practices, and calibration tips — delivered to your inbox.
AIAG average and range method — enter your measurement data and get %GRR, NDC, and pass/fail in seconds.
Open the Calculator