SlideShare a Scribd company logo
1 of 134
1
Global Procurement - Supplier Quality
Introduction to MSA
-Dinesh attri
Introduction to Measurement System Analysis
(MSA)
 Everyday our lives are being impacted by more and more data. We have become a data
driven society.
 In business and industry, we are using data in more ways than ever before.
 Today manufacturing companies gather massive amounts of information through
measurement and inspection. When this measurement data is being used to make decisions
regarding the process and the business in general it is vital that the data is accurate. If there
are errors in our measurement system we will be making decisions based on incorrect data.
We could be making incorrect decisions or producing non-conforming parts. A properly
planned and executed Measurement System Analysis (MSA) can help build a strong
foundation for any data based decision making process.
2
What is Measurement System Analysis (MSA)
 MSA is defined as an experimental and mathematical method of
determining the amount of variation that exists within a measurement
process. Variation in the measurement process can directly contribute
to our overall process variability. MSA is used to certify the
measurement system for use by evaluating the system’s accuracy,
precision and stability.
3
What is a Measurement System?
 Before we dive further into MSA, we should review the definition of a
measurement system and some of the common sources of variation.
A measurement system has been described as a system of related
measures that enables the quantification of particular characteristics.
It can also include a collection of gages, fixtures, software and
personnel required to validate a particular unit of measure or make an
assessment of the feature or characteristic being measured.
4
What is a Measurement System?
Variation
Think of Measurement as a
Process
5
What is a Measurement System?
Measurement
The assignment of numbers to material things to represent the relationships
among them with respect to particular properties.
C. Eisenhart (1963)
6
What is a Measurement System?
 The sources of variation in a measurement process can include the following:
 Process – test method, specification
 Personnel – the operators, their skill level, training, etc.
 Tools / Equipment – gages, fixtures, test equipment used and their associated
calibration systems
 Items to be measured – the part or material samples measured, the sampling
plan, etc.
 Environmental factors – temperature, humidity, etc.
7
What is a Measurement System?
 All of these possible sources of variation should be considered during
Measurement System Analysis. Evaluation of a measurement system
should include the use of specific quality tools to identify the most
likely source of variation. Most MSA activities examine two primary
sources of variation, the parts and the measurement of those parts.
The sum of these two values represents the total variation in a
measurement system.
8
Why Perform Measurement System Analysis (MSA)
 An effective MSA process can help assure that the data being
collected is accurate and the system of collecting the data is
appropriate to the process.
 Good reliable data can prevent wasted time, labor and scrap in a
manufacturing process.
9
Why Perform Measurement System Analysis (MSA)
 A major manufacturing company began receiving calls from several of their
customers reporting non-compliant materials received at their facilities sites. The
parts were not properly snapping together to form an even surface or would not
lock in place.
 The process was audited and found that the parts were being produced out of
spec. The operator was following the inspection plan and using the assigned
gages for the inspection. The problem was that the gage did not have adequate
resolution to detect the non-conforming parts.
 An ineffective measurement system can allow bad parts to be accepted and good
parts to be rejected, resulting in dissatisfied customers and excessive scrap. MSA
could have prevented the problem and assured that accurate useful data was
being collected..
Example
10
How to Perform Measurement System Analysis (MSA)
 MSA is a collection of experiments and analysis performed to evaluate a
measurement system’s capability, performance and amount of uncertainty
regarding the values measured. We should review the measurement data
being collected, the methods and tools used to collect and record the data.
 Our goal is to quantify the effectiveness of the measurement system, analyze
the variation in the data and determine its likely source. We need to evaluate
the quality of the data being collected in regards to location and width
variation. Data collected should be evaluated for bias, stability and linearity.
11
How to Perform Measurement System Analysis (MSA)
 During an MSA activity, the amount of measurement uncertainty must
be evaluated for each type of gage or measurement tool defined
within the process Control Plans.
 Each tool should have the correct level of discrimination and
resolution to obtain useful data. The process, the tools being used
(gages, fixtures, instruments, etc.) and the operators are evaluated for
proper definition, accuracy, precision, repeatability and reproducibility.
12
How to Perform Measurement System Analysis (MSA)
Data Classifications
 Prior to analyzing the data and or the gages, tools or fixtures, we
must determine the type of data being collected. The data could be
attribute data or variable data.
 Attribute data is classified into specific values where variable or
continuous data can have an infinite number of values.
13
How to Perform Measurement System Analysis (MSA)
The Master Sample
 To perform a study, you should first obtain a sample and establish the
reference value compared to a traceable standard. Some processes
will already have “master samples” established for the high and low
end of the expected measurement specification.
14
How to Perform Measurement System Analysis (MSA)
The Gage R&R Study
 For gages or instruments used to collect variable continuous
data, Gage Repeatability and Reproducibility (Gage R & R) can be
performed to evaluate the level of uncertainty within a measurement
system.
15
How to Perform Measurement System Analysis (MSA)
 To perform a Gage R & R, first select the gage to be evaluated.
 Then perform the following steps:
 Obtain at least 10 random samples of parts manufactured during a regular production run
 Choose three operators that regularly perform the particular inspection
 Have each of the operators measure the sample parts and record the data
 Repeat the measurement process three times with each operator using the same parts
 Calculate the average (mean) readings and the range of the trial averages for each of the operators
 Calculate the difference of each operator’s averages, average range and the range of
measurements for each sample part used in the study
 Calculate repeatability to determine the amount of equipment variation
 Calculate reproducibility to determine the amount of variation introduced by the operators
 Calculate the variation in the parts and total variation percentages
16
How to Perform Measurement System Analysis (MSA)
 The resulting Gage R & R percentage is used as a basis for accepting the gage.
Guidelines for making the determination are found below:
 The measurement system is acceptable if the Gage R & R score falls below 10%
 The measurement system may be determined acceptable depending upon the relative
importance of the application or other factors if the Gage R & R falls between 10% to
20%
 Any measurement system with Gage R & R greater than 30% requires action to
improve
 Any actions identified to improve the measurement system should be evaluated
for effectiveness
17
How to Perform Measurement System Analysis (MSA)
 When interpreting the results of a Gage R & R, perform a comparison study
of the repeatability and reproducibility values.
 If the repeatability value is large in comparison to the reproducibility value, it
would indicate a possible issue with the gage used for the study.
 The gage may need to be replaced or re-calibrated.
 Adversely, if the reproducibility value is large in comparison with the
repeatability value, it would indicate the variation is operator related.
 The operator may need additional training on the proper use of the gage or a
fixture may be required to assist the operator in using the gage.
18
How to Perform Measurement System Analysis (MSA)
 Gage R & R studies shall be conducted under any of the following
circumstances:
 Whenever a new or different measurement system is introduced
 Following any improvement activities
 When a different type of measurement system is introduced
 Following any improvement activities performed on the current
measurement system due to the results of a previous Gage R & R study
 Annually in alignment with set calibration schedule of the gage
19
How to Perform Measurement System Analysis (MSA)
 Attribute Gage R & R
 Attribute measurement systems can be analyzed using a similar method. Measurement
uncertainty of attribute gages shall be calculated using shorter method as below:
 Determine the gage to be studied
 Obtain 10 random samples from a regular production run
 Select 2 different operators who perform the particular inspection activity regularly
 Have the operators perform the inspection two times for each of the sample parts and record
the data
 Next, calculate the kappa value.
 When the kappa value is greater than 0.6, the gage is deemed acceptable
 If not, the gage may need to be replaced or calibrated
20
How to Perform Measurement System Analysis (MSA)
 Attribute Gage R & R
 The attribute gage study should be performed based on the same criteria listed
previously for the Gage R & R study.
 During MSA, the Gage R&R or the attribute gage study should be completed on
each of the gages, instruments or fixtures used in the measurement system. The
results should be documented and stored in a database for future reference. It
may be required for a PPAP submission to the customer.
 Furthermore, if any issues should arise, a new study can be performed on the
gage and the results compared to the previous data to determine if a change has
occurred. A properly performed MSA can have a dramatic influence on the quality
of data being collected and product quality.
21
Key terms and definitions
 Attribute data – Data that can be counted for recording and analysis (sometimes referred to as go/ no go data)
 Variable data – Data that can be measured; data that has a value that can vary from one sample to the next;
continuous variable data can have an infinite number of values
 Bias – Difference between the average or mean observed value and the target value
 Stability – A change in the measurement bias over a period of time
 A stable process would be considered in “statistical control”
 Linearity – A change in bias value within the range of normal process operation
 Resolution – Smallest unit of measure of a selected tool gage or instrument; the sensitivity of the measurement
system to process variation for a particular characteristic being measured
22
Key terms and definitions
 Accuracy – The closeness of the data to the target or exact value or to an accepted
reference value
 Precision – How close a set of measurements are to each other
 Repeatability – A measure of the effectiveness of the tool being used; the variation of
measurements obtained by a single operator using the same tool to measure the same
characteristic
 Reproducibility – A measure of the operator variation; the variation in a set of data collected
by different operators using the same tool to measure the same part characteristic
23
Key terms and definitions
 Accuracy – The closeness of the data to the target or exact value or to an accepted
reference value
 Precision – How close a set of measurements are to each other
 Repeatability – A measure of the effectiveness of the tool being used; the variation of
measurements obtained by a single operator using the same tool to measure the same
characteristic
 Reproducibility – A measure of the operator variation; the variation in a set of data collected
by different operators using the same tool to measure the same part characteristic
24
Measurement Systems
Analysis
25
Measurement Systems Analysis
 Basic Concepts of Measurement Systems
 A Process
 Statistics and the Analysis of Measurement Systems
 Conducting a Measurement Systems Analysis
 ISO - TC 69 is the Statistics Group
 Ensures high ‘Data Quality’ (Think of Bias)
26
Course Focus & Flow
 Measurement as a Process
 Mechanical Aspects (vs Destructive)
 Piece part
 Continuous (fabric)
 Features of a Measurement System
 Methods of Analysis
 Gauge R&R Studies
 Special Gauging Situations
 Go/No-Go
 Destructive Tests
27
Place Timeline Here
28
The Target & Goal
Prototype
Pre-Launch
Production
USL
LSL
Continuous Improvement 29
Key Words
 Discrimination
Ability to tell things apart
 Bias [per AIAG] (Accuracy)
 Repeatability [per AIAG] (Precision)
 Reproducibility
 Linearity
 Stability
30
Terminology
 Error ≠ Mistake
 Error ≠ Uncertainty
 Percentage Error ≠ Percentage Uncertainty
 Accuracy ≠ Precision
31
Measurement Uncertainty
 Different conventions are used to report measurement uncertainty.
 What does ±5 mean in m = 75 ±5?
 Estimated Standard Deviation: 
 Estimated Standard Error: m = /√N
 Expanded Uncertainty of ± 2 or 3
Sometimes ± 1 (Why?)
 95% or 99% Confidence Interval
 Standard Uncertainty: u
 Combined Standard Uncertainty: uc
32
Measurement Uncertainty
 Typical Reports
 Physici
33
Measurement as a Process
Basic Concepts
 Components of the Measurement System
 Requirements of a Measurement System
 Factors Affecting a Measurement System
 Characteristics of a Measurement System
Features (Qualities) of a Measurement Number
 Units (Scale)
 Accuracy
 Precision (Consistency or Repeatability)
 Resolution (Reproducibility)
34
Measurement Related Systems
Typical Experiences with
Measurement Systems
35
Basic Concepts
 Every Process Produces a “Product”
 Every Product Possesses Qualities (Features)
 Every Quality Feature Can Be Measured
 Total Variation
= Product Variation + Measurement Variation
 Some Variation Inherent in System Design
 Some Variation is Due to a Faulty Performance of the System(s)
36
The Measurement Process
What is the ‘Product’ of the Measurement Process?
What are the Features or Qualities of this Product?
How Can We Measure Those Features?
37
Measurement Systems Components
 Material to be Inspected
Piece
Continuous
 Characteristic to be Measured
 Collecting and Preparing Specimens
 Type and Scale of Measurement
 Instrument or Test Set
 Inspector or Technician
AIAG calls these ‘Appraiser’
 Conditions of Use
38
Where Does It Start?
During the Design (APQP) Stage:
The engineer responsible for determining inspections and tests, and for
specifying appropriate equipment should be well versed in measurement
systems. The Calibration folks should be part of the process as a part of a
cross-functional team.
Variability chosen instrument must be small when compared with:
Process Variability
Specification Limits
39
Typical Progression
Determine ‘Critical’
Characteristic
Determine What
Equipment is Already
Available
Determine Required
Resolution
Consideration of the Entire
Measurement System for
the Characteristic
(Variables)
Cross-Functional
Product Engineer
Product Engineer
Metrology
How will the data be
used?
40
Measurement Systems Variables
Measurement
Instrument Environment
Material Inspector Methods
Sample
Preparation
Sample
Collection
Parallax
Reproducibility
Training
Practice
Ergonomics
Test Method
Workmanship
Samples
Standards
Discrimination
Repeatability
Bias
Calibration
Linearity
Vibration
Lighting
Temperature
Humidity
These are some of the variables in a measurement
system. What others can you think of?
Fixture
Eyesight
Air Pressure
Air Movement
Fatigue
41
Determining What To Measure
 Voice of the Customer
You Must Convert to Technical Features
 Technical Features
 Failure Modes Analysis
 Control Plan
Convert To
External
Requirements
Internal
Requirements
42
Voice of the Customer
 External and Internal
Customers
 Stated vs Real and
Perceived Needs
 Cultural Needs
 Unintended Uses
 Functional Needs vs.
Technical Features
Customer may specify causes
rather than output
43
Convert to Technical Features
 Agreed upon Measure(s)
 Related to Functional Needs
 Understandable
 Uniform Interpretation
 Broad Application
 Economical
 Compatible
 Basis for Decisions
Y
Z
Technical Feature
Functional Need
44
Failure Modes Analysis
 Design FMEA
 Process FMEA
 Identify Key Features
 Identify Control Needs
Critical Features are Defined Here!
45
Automotive FMEA
Process Failure Mode And Effects Analysis Low - High
Process: Outside Suppliers Affected: Engineer: 1 - 10
Primary Process Responsibility: Model Year/Vehicle(s): Part Number:
Other Div. Or People Involved: Scheduled Production Released: PFMEA Date: Rev.
Approvals: Quality Assurance Manager Quality Assurance Engineer
Operations Manager Senior Advisor
Part Name
Operation
Number Process Function
Potential Failure
Mode
Potential Effects Of
Failure Potential Cause Of Failure Current Controls
Occured
Se
verity
De
te
ctio
n
RPN
Recommended
Actions And
Status
Actions
Taken
Occured
Se
verity
De
te
ctio
n
RPN
Responsible
Activity
SIR Take T PPE Wrong Material Fragmented Container Insufficient Supplier Control Material Certif
ication 1 9 2 18
Container Material Held In Unpredictable Deployment Improper Handling Required With Each
1 Storage Area Misidentif
ied Material Shipment
Release Verification
Out Of Spec Fragmented Container Supplier Process Control Periodic Audit Of 3 10 3 90
Material Unpredictable Deployment Supplier Material
Contaminated Fragmented Container Open Boxes Visual Inspection 1 9 7 63
Material Unpredictable Deployment
Material Fragmented Container Engineering Change Release Verification 1 10 7 70
Composition Unpredictable Deployment Supplier Change Green "OK" Tag
Change Customer Notification
2 Move To Unreleased Fragmentation Untrained LTO Check For Green "OK" 5 10 1 50
Approved Untrained Personnel Tag At Press
Storage Trace Card
Check List
Training
Leading to MSA. Critical features are determined by the FMEA
(RPN indicators) and put into the Control Plan.
Control Plan / Flow Diagram
 Inspection Points
 Inspection Frequency
 Instrument
 Measurement Scale
 Sample Preparation
 Inspection/Test Method
 Inspector (who?)
 Method of Analysis
47
GM Process Flow Chart
Process Flow Diagram Approved By:
Part Number: Date: 4/5/93 QA Manager
Part Description: Rev. : C Operations Manager
Prepared By: Senior Advisor
QA Engineer
Step
Fabrication
Move
Store
Inspect
Operation Description Item # Key Product Characteristic Item # Key Control Characteristic
1 Move "OK" Vinyl Material 1.0 Material Specs 1.0 Material Certification Tag
From Storage Area and
Load Into Press.
2 Auto Injection Mold Cover 2.0 Tearstrip In Cover 2.1 Tool Setup
In Tool # 2.2 Machine Setup
3.0 Hole Diameter In Cover 2.1 Tool Setup
2.2 Machine Setup
4.0 Flange Thickness In Cover 2.1 Tool Setup
2.2 Machine Setup
5.0 Pressure Control Protrusions 2.1 Tool Setup
Height 2.2 Machine Setup
3 Visually Inspect Cover 6.0 Pressure Control Protrusions 2.1 Tool Setup
Filled Out 2.2 Machine Setup
Standard Control Plan Example
Control Plan Number Key Contact / Phone Date (Orig.) Date (Rev.)
Part No./ Latest Change No. Core Team Customer Engineering Approval/Date
Part Name/Description Supplier/Plant Apoproval/Date Customer Quality Approval/Date
Supplier/Plant Supplier Code Other Approval/date (If Req'd) Other Approval/date (If Req'd)
Characteristics Methods
Part/
Process
Number
Process Name/
Operation
Description
Machine,
Device,
Jig, Tools
for Mfg. No. Product Process
Special
Char.
Class
Product/
Process
Spec/
Tolerance
Evaluation
Measurement
Technique Size
Frequ-
ency
Control
Method
Reaction
Plan
This form is on course disk
49
Ford’s Dimensional Control Plan (DCP)
50
Measurement as a System
 Choosing the Right Instrument
 Instrument Calibration Needs
 Standards or Masters Needed
 Accuracy and Precision
 Measurement Practices
 Where
 How Many Places
 Reported Figures
 Significant Figures Rule
 2 Action Figures
 Rule of 10
 Individuals, Averages, High-Lows
51
Measurement Error
Measured Value (y)
=
True Value (x) + Measurement Error
Deming says there is no
such thing as a ‘True’ Value. Consistent (linear)?
52
Sources of Measurement Error
 Sensitivity (Threshold)
Chemical Indicators
 Discrimination
 Precision (Repeatability)
 Accuracy (Bias)
 Damage
 Differences in use by Inspector (Reproducibility)
Training Issues
 Differences Among Instruments and Fixtures
 Differences Among Methods of Use
 Differences Due to Environment
53
Types of Measurement Scales
 Variables
 Can be measured on a continuous scale
 Defined, standard Units of Measurement
 Attributes
 No scale
 Derived ‘Unit of Measurement’
 Can be observed or counted
 Either present or not
 Needs large sample size because of low information content
54
How We Get Data
Inspection
Measurement
Test
Includes Sensory (e.g..: look,
touch, smell…etc)
Magnitude of Quality
55
Operational Definitions
 Is the container Round?
 Is your software Accurate?
 Is the computer screen Clean?
 Is the truck On Time?
56
Different Method = Different Results
In Spec
Out of
Spec
Method 1
Method 2
57
Measurement System Variability
 Small with respect to Process Variation
 Small with respect to Specified Requirements
 Must be in Statistical Control
Measurement IS a Process!
Free of Assignable Causes of variation
58
Studying the Measurement System
 Environmental Factors
 Human Factors
 System Features
 Measurement Studies
59
Environmental Factors
 Temperature
 Humidity
 Vibration
 Lighting
 Corrosion
 Wear
 Contaminants
Oil & Grease
Aerosols
Where is the study performed?
1. Lab?
2. Where used?
3. Both?
60
Human Factors
 Training
 Skills
 Fatigue
 Boredom
 Eyesight
 Comfort
 Complexity of Part
 Speed of Inspection (parts per hour)
 Misunderstood Instructions
61
Human Measurement Errors
 Sources of Errors
Inadvertent Errors
 Attentiveness
 Random
 Good Mistake-Proofing Target
Technique Errors
 Consistent
Wilful Errors (Bad mood)
 Error Types (Can be machine or human)
Type I - Alpha Errors [ risk]
Type II - Beta Errors [ risk]
Accept
Reject
Good Bad
OK!
OK!
alpha
beta
Training
Issue
Process in
control, but
needs
adjustment,
False alarm
Unaware of
problem
62
Measurement System Features
 Discrimination
Ability to tell things apart
 Bias [per AIAG] (Accuracy)
 Repeatability [per AIAG] (Precision)
 Reproducibility
 Linearity
 Stability
63
Discrimination
 Readable Increments of Scale
 If Unit of Measure is too course: Process variation will be lost in
Rounding Off
 The “Rule of Ten”: Ten possible values between limits is ideal
Five Possible Values: Marginally useful
Four or Less: Inadequate Discrimination
64
Discrimination
65
Range Charts & Discrimination
Indicates Poor
Precision
66
Bias and Repeatability
Precise Imprecise
Accurate
Inaccurate
Bias
You can correct for Bias
You can NOT correct for Imprecision
67
Bias
 Difference between average of
measurements and an Agreed Upon
standard value
 Known as Accuracy
 Cannot be evaluated without a
Standard
 Adds a Consistent “Bias Factor” to
ALL measurements
 Affects all measurements in the same
way
Standard
Value
Measurement Scale
Bias
68
Causes of Bias
 Error in Master
 Worn components
 Instrument improperly calibrated
 Instrument damaged
 Instrument improperly used
 Instrument read incorrectly
 Part set incorrectly (wrong datum)
69
Bias
 Bias - The difference between the observed Average of measurements
and the master Average of the same parts using precision instruments.
(MSA Manual Glossary)
 The auditor may want evidence that the concept of bias is understood.
Remember that bias is basically an offset from ‘zero’. Bias is linked to
Stability in the sense that an instrument may be ‘zeroed’ during
calibration verification. Knowing this we deduce that the bias changes
with instrument use. This is in part the concept of Drift.
70
Bias
 I choose a caliper (resolution 0.01) for the measurement. I measure a
set of parts and derive the average.
 I take the same parts and measure them with a micrometer
(resolution 0.001). I then derive the average.
 I compare the two averages. The difference is the Bias.
71
Repeatability
 Variation among repeated
measurements
 Known as Precision
 Standard NOT required
 May add or subtract from a given
measurement
 Affects each measurement randomly
Measurement Scale
Repeatability
Margin of Error
Doesn’t address Bias
5.15 = 99%
72
Repeatability Issues
 Measurement Steps
 Sample preparation
 Setting up the instrument
 Locating on the part
 How much of the measurement process should we repeat?
73
Using Shewhart Charts I
Repeatability
74
Using Shewhart Charts II
75
Evaluating Bias & Repeatability
 Same appraiser, Same part, Same instrument
 Multiple readings (n≥10 with 20 to 40 better)
 Analysis
 Average minus Standard Value = Bias
 5.15* Standard Deviation = Repeatability
 or +/- 2.575  [99% repeatability]
 or +/- 2  [95% repeatability]
 Histogram
 Probability
AIAG
76
Repeatability Issues
 Making a measurement may involve numerous steps
 Sample preparation
 Setting up the instrument
 Locating the part, etc.
 How much of the measurement process should we repeat? How far
do we go?
77
Bias & Repeatability Histogram
Never include assignable cause errors
Linearity
 The difference in the Bias or Repeatability across the expected
operating range of the instrument.
79
Plot Biases vs. Ref. Values
Linearity = |Slope| * Process Variation = 0.1317*6.00 = 0.79
% Linearity = 100 * |Slope| = 13.17%
80
Causes of Poor Linearity
 Instrument not properly calibrated at both Upper and Lower extremes
 Error in the minimum or maximum Master
 Worn Instrument
 Instrument design characteristics
81
Reproducibility
 Variation in the averages
among different appraisers
repeatedly measuring the
same part characteristic
 Concept can also apply to
variation among different
instruments
Includes repeatability which must be accounted for.
82
Reproducibility Example
83
Calculating Reproducibility (I)
 Find the range of the appraiser averages (R0)
 Convert to Standard Deviation using d2*
(m=# of appraisers; g=# of ranges used = 1)
 Multiply by 5.15
 Subtract the portion of this due to repeatability
84
Calculating Reproducibility
People variance
Trials
Times done
85
Stability
 Variation in measurements
of a single characteristic
 On the same master
 Over an extended period
of time
 Evaluate using Shewhart charts
86
Evaluate Stability with Run Charts
87
Stability
Both gages are stable, but.....
88
Importance of Stability
 Statistical stability, combined with subject-matter knowledge, allows
predictions of process performance
 Action based on analysis of Unstable systems may increase Variation
due to ‘Tampering’
 A statistically unstable measurement system cannot provide reliable
data on the process
89
Methods of Analysis
90
Analysis Tools
 Calculations of Average and Standard Deviation
 Correlation Charts
 Multi-Vari Charts
 Box-and-Whisker Plots
 Run charts
 Shewhart charts
91
Average and Standard Deviation
92
Correlation Charts
 Describe Relationships
 Substitute measurement for desired measurement
 Actual measurement to reference value
 Inexpensive gaging method versus Expensive gaging method
 Appraiser A with appraiser B
93
Substitute Measurements
 Cannot directly measure quality
 Correlate substitute measure
 Measure substitute
 Convert to desired quality
94
Comparing Two Methods
 Two methods
 Measure parts using both
 Correlate the two
 Compare to “Line of No Bias”
 Investigate differences
Magnetic
Stripping
Line of Perfect Agreement
Line of Correlation
95
Measurements vs. Reference Data
96
Measurements vs. Reference Correlation
Disparity
97
Comparing Two Appraisers
98
Run Charts Examine Stability
99
Multiple Run Charts
More than 3 appraisers confuses things...
Multi-Vari Charts
 Displays 3 points
 Length of bar; bar-to-bar; Bar cluster to cluster
 Plot High and Low readings as Length of bar
 Each appraiser on a separate bar
 Each piece in a separate bar cluster
High Reading
Low Reading
Average Reading
101
Multi-Vari Type I
 Bar lengths are long
 Appraiser differences small
in comparison
 Piece-to-piece hard to detect
 Problem is repeatability
102
Multi-Vari Type II
 Appraiser differences are
biggest source of variation
 Bar length is small in
comparison
 Piece-to-piece hard to detect
 Problem is reproducibility
103
Multi-Vari Type III
 Piece-to-piece variation
is the biggest source of
variation
 Bar length (repeatability)
is small in comparison
 Appraiser differences
(bar-to-bar) is small in
comparison
 Ideal Pattern
104
Multi-Vari Chart Example
Normalized Data
105
Multi-Vari Chart, Joined
Look for similar pattern
106
Using Shewhart Charts
 Subgroup = Repeated measurements,, same piece
 Different Subgroups = Different pieces and/or appraisers
 Range chart shows precision (repeatability)
 Average chart “In Control” shows reproducibility
If subgroups are different appraisers
 Average chart shows discriminating power
If subgroups are different pieces
(“In Control” is BAD!)
107
Shewhart Charts
This is not a good way to plot this data
Too many lines
108
Shewhart Chart of Instrument
109
Gage R&R Studies
110
Gauge R&R Studies
 Developed by Jack Gantt
 Originally plotted on probability paper
 Revived as purely numerical calculations
 Worksheets developed by AIAG
 Renewed awareness of Measurement Systems as ‘Part of the Process’
Consider Numerical vs. Graphical Data Evaluations
111
Terms Used in R&R (I)
 n = Number of Parts [2 to 10]
Parts represent total range of process variation
Need not be “good” parts. Do NOT use consecutive pieces.
Screen for size
 a = Number of Appraisers
Each appraiser measures each part r times
Study must be by those actually using
 R - Number of trials
 Also called “m” in AIAG manual
 g = r*a [Used to find d2* in table 2, p. 29 AIAG manual]
1 Outside Low/High
1 Inside Low/High
Target
Minimum of 5.
2 to 10 To
accommodate
worksheet factors
1 2
3
4 5
112
Terms Used in R&R (II)
 R-barA = Average range for appraiser A, etc.
 R-double bar = Average of R-barA, R-barB
 Rp = Range of part averages
 XDIFF = Difference between High & Low appraiser averages
Also a range, but “R” is not used to avoid confusion
 EV = 5.15 = Equipment variation (repeatability)
 EV = 5.15 = Equipment variation (reproducibility)
 PV = Part variation
 TV = Total variation
Process Variation
113
R&R Calculations
Measurement
System Variation
Product Process
Variation
Left over
Repeatability
Remember -
Nonconsecutive
Pieces
Left over
Repeatability
114
Accumulation of Variances
115
Evaluating R&R
 %R&R=100*[R&R/TV] (Process Control)
 %R&R=100*[R&R/Tolerance] (Inspection)
 Under 10%: Measurement System Acceptable
 10% to 30%: Possibly acceptable, depending upon use, cost, etc.
 Over 30%: Needs serious improvement
116
Analysis of Variance I
 Mean squares and Sums of squares
 Ratio of variances versus expected F-ratio
 Advantages
Any experimental layout
Estimate interaction effects
 Disadvantages
Must use computer
Non-intuitive interpretation
117
Analysis of Variance II
 The n*r measurements must be done in random sequence [a good
idea anyway]
 Assumes that EV [repeatability] is normal and that EV is not
proportional to measurement [normally a fairly good assumption]
 Details beyond scope of this course
118
Special Gauging Situations
 Go/No-Go
 Destructive Testing
119
If Gauges were Perfect
120
But Repeatability Means We Never Know The Precise
Value
121
So - Actual Part Acceptance Will Look Like This:
122
The Effect of Bias on Part Acceptance
123
Go/No-Go gauges
 Treat variables like attributes
 Provide less information on the process, but...
 Are fast and inexpensive
 Cannot use for Process Control
 Can be used for Sorting purposes
124
“Short” Go/No-Go Study
 Collect 20 parts covering the entire process range
 Use two inspectors
 Gage each part twice
 Accept gauge if there is agreement on each of the 20 parts
* May reject a good measuring system
125
Destructive Tests
 Cannot make true duplicate tests
 Use interpenetrating samples
 Compare 3 averages
 Adjust using √n
126
Destructive Tests: Interpreting Samples
AIAG does not address
127
Summary
128
Measurement Variation
 Observed variation is a combination of the production process PLUS
the measurement process
 The contribution of the measurement system is often overlooked
129
Types of Measurement Variation
 Bias (Inaccuracy)
 Repeatability (Imprecision)
 Discrimination
 Linearity
 Stability
130
Measurement Systems
 Material
 Characteristic
 Sampling and Preparation
 Operational Definition of Measurement
 Instrument
 Appraiser
 Environment and Ergonomics
131
Measurement Systems Evaluation Tools
 Histograms
 Probability paper
 Run Charts
 Scatter diagrams
 Multi-Vari Charts
 Gantt “R&R” analysis
 Analysis of Variance (ANOVA)
 Shewhart “Control” Charts
132
Shewhart Charts
 Range chart shows repeatability
 X-bar limits show discriminating power
 X-double bar shows bias
(if a known standard exists)
 Average chart shows stability
(sub-groups overtime)
 Average chart shows reproducibility
(sub-groups over technicians/instruments)
133
Conclusion
 Rule of Ten
 Operating Characteristic Curve
 Special Problems
Go/No-Go Gages
Attribute Inspection
Destructive Testing
134

More Related Content

Similar to GP_Training_Introduction-to-MSA__RevAF.pptx

10. measurement system analysis (msa)
10. measurement system analysis (msa)10. measurement system analysis (msa)
10. measurement system analysis (msa)Hakeem-Ur- Rehman
 
Quality Journey- Measurement System Analysis .pdf
Quality Journey- Measurement System Analysis .pdfQuality Journey- Measurement System Analysis .pdf
Quality Journey- Measurement System Analysis .pdfNileshJajoo2
 
MSA – Planning & Conducting the MSA
MSA – Planning & Conducting the MSAMSA – Planning & Conducting the MSA
MSA – Planning & Conducting the MSAMatt Hansen
 
Measuremen Systems Analysis Training Module
Measuremen Systems Analysis Training ModuleMeasuremen Systems Analysis Training Module
Measuremen Systems Analysis Training ModuleFrank-G. Adler
 
Six Sigma Green Belt Training Part 8
Six Sigma Green Belt Training Part 8Six Sigma Green Belt Training Part 8
Six Sigma Green Belt Training Part 8Skillogic Solutions
 
Measurement System Analysis
Measurement System AnalysisMeasurement System Analysis
Measurement System AnalysisRonald Shewchuk
 
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...Gabor Szabo, CQE
 
Six sigma tools an overview
Six sigma tools  an overviewSix sigma tools  an overview
Six sigma tools an overviewKomal Kamble
 
Five costly mistakes applying spc [whitepaper]
Five costly mistakes applying spc [whitepaper]Five costly mistakes applying spc [whitepaper]
Five costly mistakes applying spc [whitepaper]Blackberry&Cross
 
02trainingmaterialformsa (1) 111030062223-phpapp02
02trainingmaterialformsa (1) 111030062223-phpapp0202trainingmaterialformsa (1) 111030062223-phpapp02
02trainingmaterialformsa (1) 111030062223-phpapp02Junelly Grace Catalan-Tecson
 
MSA presentation
MSA presentationMSA presentation
MSA presentationsanjay deo
 
Statistical Quality Control
Statistical Quality ControlStatistical Quality Control
Statistical Quality ControlMahmudul Hasan
 
090528 Miller Process Forensics Talk @ Asq
090528 Miller Process Forensics Talk @ Asq090528 Miller Process Forensics Talk @ Asq
090528 Miller Process Forensics Talk @ Asqrwmill9716
 
Sslean Validation 20070622
Sslean Validation 20070622Sslean Validation 20070622
Sslean Validation 20070622jancrielaard
 
Statistical_Quality_Control.ppt
Statistical_Quality_Control.pptStatistical_Quality_Control.ppt
Statistical_Quality_Control.pptssusera85eeb1
 
Example of quality management system
Example of quality management systemExample of quality management system
Example of quality management systemselinasimpson1701
 

Similar to GP_Training_Introduction-to-MSA__RevAF.pptx (20)

10. measurement system analysis (msa)
10. measurement system analysis (msa)10. measurement system analysis (msa)
10. measurement system analysis (msa)
 
Quality Journey- Measurement System Analysis .pdf
Quality Journey- Measurement System Analysis .pdfQuality Journey- Measurement System Analysis .pdf
Quality Journey- Measurement System Analysis .pdf
 
MSA – Planning & Conducting the MSA
MSA – Planning & Conducting the MSAMSA – Planning & Conducting the MSA
MSA – Planning & Conducting the MSA
 
Msa training
Msa trainingMsa training
Msa training
 
Measuremen Systems Analysis Training Module
Measuremen Systems Analysis Training ModuleMeasuremen Systems Analysis Training Module
Measuremen Systems Analysis Training Module
 
Six Sigma Green Belt Training Part 8
Six Sigma Green Belt Training Part 8Six Sigma Green Belt Training Part 8
Six Sigma Green Belt Training Part 8
 
Dt33726730
Dt33726730Dt33726730
Dt33726730
 
Dt33726730
Dt33726730Dt33726730
Dt33726730
 
Measurement System Analysis
Measurement System AnalysisMeasurement System Analysis
Measurement System Analysis
 
Msa presentation
Msa presentationMsa presentation
Msa presentation
 
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
Measurement Systems Analysis - Variable Gage R&R Study Metrics, Applications ...
 
Six sigma tools an overview
Six sigma tools  an overviewSix sigma tools  an overview
Six sigma tools an overview
 
Five costly mistakes applying spc [whitepaper]
Five costly mistakes applying spc [whitepaper]Five costly mistakes applying spc [whitepaper]
Five costly mistakes applying spc [whitepaper]
 
02trainingmaterialformsa (1) 111030062223-phpapp02
02trainingmaterialformsa (1) 111030062223-phpapp0202trainingmaterialformsa (1) 111030062223-phpapp02
02trainingmaterialformsa (1) 111030062223-phpapp02
 
MSA presentation
MSA presentationMSA presentation
MSA presentation
 
Statistical Quality Control
Statistical Quality ControlStatistical Quality Control
Statistical Quality Control
 
090528 Miller Process Forensics Talk @ Asq
090528 Miller Process Forensics Talk @ Asq090528 Miller Process Forensics Talk @ Asq
090528 Miller Process Forensics Talk @ Asq
 
Sslean Validation 20070622
Sslean Validation 20070622Sslean Validation 20070622
Sslean Validation 20070622
 
Statistical_Quality_Control.ppt
Statistical_Quality_Control.pptStatistical_Quality_Control.ppt
Statistical_Quality_Control.ppt
 
Example of quality management system
Example of quality management systemExample of quality management system
Example of quality management system
 

Recently uploaded

Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17Celine George
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Krashi Coaching
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxOH TEIK BIN
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityGeoBlogs
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxNirmalaLoungPoorunde1
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxpboyjonauth
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Sapana Sha
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxAvyJaneVismanos
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxthorishapillay1
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...Marc Dusseiller Dusjagr
 
Painted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaPainted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaVirag Sontakke
 
Class 11 Legal Studies Ch-1 Concept of State .pdf
Class 11 Legal Studies Ch-1 Concept of State .pdfClass 11 Legal Studies Ch-1 Concept of State .pdf
Class 11 Legal Studies Ch-1 Concept of State .pdfakmcokerachita
 
भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,Virag Sontakke
 
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdfBASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdfSoniaTolstoy
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsanshu789521
 
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfSumit Tiwari
 

Recently uploaded (20)

Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17How to Configure Email Server in Odoo 17
How to Configure Email Server in Odoo 17
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
 
Solving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptxSolving Puzzles Benefits Everyone (English).pptx
Solving Puzzles Benefits Everyone (English).pptx
 
Paris 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activityParis 2024 Olympic Geographies - an activity
Paris 2024 Olympic Geographies - an activity
 
Employee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptxEmployee wellbeing at the workplace.pptx
Employee wellbeing at the workplace.pptx
 
Introduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptxIntroduction to AI in Higher Education_draft.pptx
Introduction to AI in Higher Education_draft.pptx
 
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111Call Girls in Dwarka Mor Delhi Contact Us 9654467111
Call Girls in Dwarka Mor Delhi Contact Us 9654467111
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
Final demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptxFinal demo Grade 9 for demo Plan dessert.pptx
Final demo Grade 9 for demo Plan dessert.pptx
 
Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
Proudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptxProudly South Africa powerpoint Thorisha.pptx
Proudly South Africa powerpoint Thorisha.pptx
 
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
“Oh GOSH! Reflecting on Hackteria's Collaborative Practices in a Global Do-It...
 
Painted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of IndiaPainted Grey Ware.pptx, PGW Culture of India
Painted Grey Ware.pptx, PGW Culture of India
 
Class 11 Legal Studies Ch-1 Concept of State .pdf
Class 11 Legal Studies Ch-1 Concept of State .pdfClass 11 Legal Studies Ch-1 Concept of State .pdf
Class 11 Legal Studies Ch-1 Concept of State .pdf
 
भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,भारत-रोम व्यापार.pptx, Indo-Roman Trade,
भारत-रोम व्यापार.pptx, Indo-Roman Trade,
 
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdfBASLIQ CURRENT LOOKBOOK  LOOKBOOK(1) (1).pdf
BASLIQ CURRENT LOOKBOOK LOOKBOOK(1) (1).pdf
 
Presiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha electionsPresiding Officer Training module 2024 lok sabha elections
Presiding Officer Training module 2024 lok sabha elections
 
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdfEnzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
Enzyme, Pharmaceutical Aids, Miscellaneous Last Part of Chapter no 5th.pdf
 
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝Model Call Girl in Bikash Puri  Delhi reach out to us at 🔝9953056974🔝
Model Call Girl in Bikash Puri Delhi reach out to us at 🔝9953056974🔝
 

GP_Training_Introduction-to-MSA__RevAF.pptx

  • 1. 1 Global Procurement - Supplier Quality Introduction to MSA -Dinesh attri
  • 2. Introduction to Measurement System Analysis (MSA)  Everyday our lives are being impacted by more and more data. We have become a data driven society.  In business and industry, we are using data in more ways than ever before.  Today manufacturing companies gather massive amounts of information through measurement and inspection. When this measurement data is being used to make decisions regarding the process and the business in general it is vital that the data is accurate. If there are errors in our measurement system we will be making decisions based on incorrect data. We could be making incorrect decisions or producing non-conforming parts. A properly planned and executed Measurement System Analysis (MSA) can help build a strong foundation for any data based decision making process. 2
  • 3. What is Measurement System Analysis (MSA)  MSA is defined as an experimental and mathematical method of determining the amount of variation that exists within a measurement process. Variation in the measurement process can directly contribute to our overall process variability. MSA is used to certify the measurement system for use by evaluating the system’s accuracy, precision and stability. 3
  • 4. What is a Measurement System?  Before we dive further into MSA, we should review the definition of a measurement system and some of the common sources of variation. A measurement system has been described as a system of related measures that enables the quantification of particular characteristics. It can also include a collection of gages, fixtures, software and personnel required to validate a particular unit of measure or make an assessment of the feature or characteristic being measured. 4
  • 5. What is a Measurement System? Variation Think of Measurement as a Process 5
  • 6. What is a Measurement System? Measurement The assignment of numbers to material things to represent the relationships among them with respect to particular properties. C. Eisenhart (1963) 6
  • 7. What is a Measurement System?  The sources of variation in a measurement process can include the following:  Process – test method, specification  Personnel – the operators, their skill level, training, etc.  Tools / Equipment – gages, fixtures, test equipment used and their associated calibration systems  Items to be measured – the part or material samples measured, the sampling plan, etc.  Environmental factors – temperature, humidity, etc. 7
  • 8. What is a Measurement System?  All of these possible sources of variation should be considered during Measurement System Analysis. Evaluation of a measurement system should include the use of specific quality tools to identify the most likely source of variation. Most MSA activities examine two primary sources of variation, the parts and the measurement of those parts. The sum of these two values represents the total variation in a measurement system. 8
  • 9. Why Perform Measurement System Analysis (MSA)  An effective MSA process can help assure that the data being collected is accurate and the system of collecting the data is appropriate to the process.  Good reliable data can prevent wasted time, labor and scrap in a manufacturing process. 9
  • 10. Why Perform Measurement System Analysis (MSA)  A major manufacturing company began receiving calls from several of their customers reporting non-compliant materials received at their facilities sites. The parts were not properly snapping together to form an even surface or would not lock in place.  The process was audited and found that the parts were being produced out of spec. The operator was following the inspection plan and using the assigned gages for the inspection. The problem was that the gage did not have adequate resolution to detect the non-conforming parts.  An ineffective measurement system can allow bad parts to be accepted and good parts to be rejected, resulting in dissatisfied customers and excessive scrap. MSA could have prevented the problem and assured that accurate useful data was being collected.. Example 10
  • 11. How to Perform Measurement System Analysis (MSA)  MSA is a collection of experiments and analysis performed to evaluate a measurement system’s capability, performance and amount of uncertainty regarding the values measured. We should review the measurement data being collected, the methods and tools used to collect and record the data.  Our goal is to quantify the effectiveness of the measurement system, analyze the variation in the data and determine its likely source. We need to evaluate the quality of the data being collected in regards to location and width variation. Data collected should be evaluated for bias, stability and linearity. 11
  • 12. How to Perform Measurement System Analysis (MSA)  During an MSA activity, the amount of measurement uncertainty must be evaluated for each type of gage or measurement tool defined within the process Control Plans.  Each tool should have the correct level of discrimination and resolution to obtain useful data. The process, the tools being used (gages, fixtures, instruments, etc.) and the operators are evaluated for proper definition, accuracy, precision, repeatability and reproducibility. 12
  • 13. How to Perform Measurement System Analysis (MSA) Data Classifications  Prior to analyzing the data and or the gages, tools or fixtures, we must determine the type of data being collected. The data could be attribute data or variable data.  Attribute data is classified into specific values where variable or continuous data can have an infinite number of values. 13
  • 14. How to Perform Measurement System Analysis (MSA) The Master Sample  To perform a study, you should first obtain a sample and establish the reference value compared to a traceable standard. Some processes will already have “master samples” established for the high and low end of the expected measurement specification. 14
  • 15. How to Perform Measurement System Analysis (MSA) The Gage R&R Study  For gages or instruments used to collect variable continuous data, Gage Repeatability and Reproducibility (Gage R & R) can be performed to evaluate the level of uncertainty within a measurement system. 15
  • 16. How to Perform Measurement System Analysis (MSA)  To perform a Gage R & R, first select the gage to be evaluated.  Then perform the following steps:  Obtain at least 10 random samples of parts manufactured during a regular production run  Choose three operators that regularly perform the particular inspection  Have each of the operators measure the sample parts and record the data  Repeat the measurement process three times with each operator using the same parts  Calculate the average (mean) readings and the range of the trial averages for each of the operators  Calculate the difference of each operator’s averages, average range and the range of measurements for each sample part used in the study  Calculate repeatability to determine the amount of equipment variation  Calculate reproducibility to determine the amount of variation introduced by the operators  Calculate the variation in the parts and total variation percentages 16
  • 17. How to Perform Measurement System Analysis (MSA)  The resulting Gage R & R percentage is used as a basis for accepting the gage. Guidelines for making the determination are found below:  The measurement system is acceptable if the Gage R & R score falls below 10%  The measurement system may be determined acceptable depending upon the relative importance of the application or other factors if the Gage R & R falls between 10% to 20%  Any measurement system with Gage R & R greater than 30% requires action to improve  Any actions identified to improve the measurement system should be evaluated for effectiveness 17
  • 18. How to Perform Measurement System Analysis (MSA)  When interpreting the results of a Gage R & R, perform a comparison study of the repeatability and reproducibility values.  If the repeatability value is large in comparison to the reproducibility value, it would indicate a possible issue with the gage used for the study.  The gage may need to be replaced or re-calibrated.  Adversely, if the reproducibility value is large in comparison with the repeatability value, it would indicate the variation is operator related.  The operator may need additional training on the proper use of the gage or a fixture may be required to assist the operator in using the gage. 18
  • 19. How to Perform Measurement System Analysis (MSA)  Gage R & R studies shall be conducted under any of the following circumstances:  Whenever a new or different measurement system is introduced  Following any improvement activities  When a different type of measurement system is introduced  Following any improvement activities performed on the current measurement system due to the results of a previous Gage R & R study  Annually in alignment with set calibration schedule of the gage 19
  • 20. How to Perform Measurement System Analysis (MSA)  Attribute Gage R & R  Attribute measurement systems can be analyzed using a similar method. Measurement uncertainty of attribute gages shall be calculated using shorter method as below:  Determine the gage to be studied  Obtain 10 random samples from a regular production run  Select 2 different operators who perform the particular inspection activity regularly  Have the operators perform the inspection two times for each of the sample parts and record the data  Next, calculate the kappa value.  When the kappa value is greater than 0.6, the gage is deemed acceptable  If not, the gage may need to be replaced or calibrated 20
  • 21. How to Perform Measurement System Analysis (MSA)  Attribute Gage R & R  The attribute gage study should be performed based on the same criteria listed previously for the Gage R & R study.  During MSA, the Gage R&R or the attribute gage study should be completed on each of the gages, instruments or fixtures used in the measurement system. The results should be documented and stored in a database for future reference. It may be required for a PPAP submission to the customer.  Furthermore, if any issues should arise, a new study can be performed on the gage and the results compared to the previous data to determine if a change has occurred. A properly performed MSA can have a dramatic influence on the quality of data being collected and product quality. 21
  • 22. Key terms and definitions  Attribute data – Data that can be counted for recording and analysis (sometimes referred to as go/ no go data)  Variable data – Data that can be measured; data that has a value that can vary from one sample to the next; continuous variable data can have an infinite number of values  Bias – Difference between the average or mean observed value and the target value  Stability – A change in the measurement bias over a period of time  A stable process would be considered in “statistical control”  Linearity – A change in bias value within the range of normal process operation  Resolution – Smallest unit of measure of a selected tool gage or instrument; the sensitivity of the measurement system to process variation for a particular characteristic being measured 22
  • 23. Key terms and definitions  Accuracy – The closeness of the data to the target or exact value or to an accepted reference value  Precision – How close a set of measurements are to each other  Repeatability – A measure of the effectiveness of the tool being used; the variation of measurements obtained by a single operator using the same tool to measure the same characteristic  Reproducibility – A measure of the operator variation; the variation in a set of data collected by different operators using the same tool to measure the same part characteristic 23
  • 24. Key terms and definitions  Accuracy – The closeness of the data to the target or exact value or to an accepted reference value  Precision – How close a set of measurements are to each other  Repeatability – A measure of the effectiveness of the tool being used; the variation of measurements obtained by a single operator using the same tool to measure the same characteristic  Reproducibility – A measure of the operator variation; the variation in a set of data collected by different operators using the same tool to measure the same part characteristic 24
  • 26. Measurement Systems Analysis  Basic Concepts of Measurement Systems  A Process  Statistics and the Analysis of Measurement Systems  Conducting a Measurement Systems Analysis  ISO - TC 69 is the Statistics Group  Ensures high ‘Data Quality’ (Think of Bias) 26
  • 27. Course Focus & Flow  Measurement as a Process  Mechanical Aspects (vs Destructive)  Piece part  Continuous (fabric)  Features of a Measurement System  Methods of Analysis  Gauge R&R Studies  Special Gauging Situations  Go/No-Go  Destructive Tests 27
  • 29. The Target & Goal Prototype Pre-Launch Production USL LSL Continuous Improvement 29
  • 30. Key Words  Discrimination Ability to tell things apart  Bias [per AIAG] (Accuracy)  Repeatability [per AIAG] (Precision)  Reproducibility  Linearity  Stability 30
  • 31. Terminology  Error ≠ Mistake  Error ≠ Uncertainty  Percentage Error ≠ Percentage Uncertainty  Accuracy ≠ Precision 31
  • 32. Measurement Uncertainty  Different conventions are used to report measurement uncertainty.  What does ±5 mean in m = 75 ±5?  Estimated Standard Deviation:   Estimated Standard Error: m = /√N  Expanded Uncertainty of ± 2 or 3 Sometimes ± 1 (Why?)  95% or 99% Confidence Interval  Standard Uncertainty: u  Combined Standard Uncertainty: uc 32
  • 33. Measurement Uncertainty  Typical Reports  Physici 33
  • 34. Measurement as a Process Basic Concepts  Components of the Measurement System  Requirements of a Measurement System  Factors Affecting a Measurement System  Characteristics of a Measurement System Features (Qualities) of a Measurement Number  Units (Scale)  Accuracy  Precision (Consistency or Repeatability)  Resolution (Reproducibility) 34
  • 35. Measurement Related Systems Typical Experiences with Measurement Systems 35
  • 36. Basic Concepts  Every Process Produces a “Product”  Every Product Possesses Qualities (Features)  Every Quality Feature Can Be Measured  Total Variation = Product Variation + Measurement Variation  Some Variation Inherent in System Design  Some Variation is Due to a Faulty Performance of the System(s) 36
  • 37. The Measurement Process What is the ‘Product’ of the Measurement Process? What are the Features or Qualities of this Product? How Can We Measure Those Features? 37
  • 38. Measurement Systems Components  Material to be Inspected Piece Continuous  Characteristic to be Measured  Collecting and Preparing Specimens  Type and Scale of Measurement  Instrument or Test Set  Inspector or Technician AIAG calls these ‘Appraiser’  Conditions of Use 38
  • 39. Where Does It Start? During the Design (APQP) Stage: The engineer responsible for determining inspections and tests, and for specifying appropriate equipment should be well versed in measurement systems. The Calibration folks should be part of the process as a part of a cross-functional team. Variability chosen instrument must be small when compared with: Process Variability Specification Limits 39
  • 40. Typical Progression Determine ‘Critical’ Characteristic Determine What Equipment is Already Available Determine Required Resolution Consideration of the Entire Measurement System for the Characteristic (Variables) Cross-Functional Product Engineer Product Engineer Metrology How will the data be used? 40
  • 41. Measurement Systems Variables Measurement Instrument Environment Material Inspector Methods Sample Preparation Sample Collection Parallax Reproducibility Training Practice Ergonomics Test Method Workmanship Samples Standards Discrimination Repeatability Bias Calibration Linearity Vibration Lighting Temperature Humidity These are some of the variables in a measurement system. What others can you think of? Fixture Eyesight Air Pressure Air Movement Fatigue 41
  • 42. Determining What To Measure  Voice of the Customer You Must Convert to Technical Features  Technical Features  Failure Modes Analysis  Control Plan Convert To External Requirements Internal Requirements 42
  • 43. Voice of the Customer  External and Internal Customers  Stated vs Real and Perceived Needs  Cultural Needs  Unintended Uses  Functional Needs vs. Technical Features Customer may specify causes rather than output 43
  • 44. Convert to Technical Features  Agreed upon Measure(s)  Related to Functional Needs  Understandable  Uniform Interpretation  Broad Application  Economical  Compatible  Basis for Decisions Y Z Technical Feature Functional Need 44
  • 45. Failure Modes Analysis  Design FMEA  Process FMEA  Identify Key Features  Identify Control Needs Critical Features are Defined Here! 45
  • 46. Automotive FMEA Process Failure Mode And Effects Analysis Low - High Process: Outside Suppliers Affected: Engineer: 1 - 10 Primary Process Responsibility: Model Year/Vehicle(s): Part Number: Other Div. Or People Involved: Scheduled Production Released: PFMEA Date: Rev. Approvals: Quality Assurance Manager Quality Assurance Engineer Operations Manager Senior Advisor Part Name Operation Number Process Function Potential Failure Mode Potential Effects Of Failure Potential Cause Of Failure Current Controls Occured Se verity De te ctio n RPN Recommended Actions And Status Actions Taken Occured Se verity De te ctio n RPN Responsible Activity SIR Take T PPE Wrong Material Fragmented Container Insufficient Supplier Control Material Certif ication 1 9 2 18 Container Material Held In Unpredictable Deployment Improper Handling Required With Each 1 Storage Area Misidentif ied Material Shipment Release Verification Out Of Spec Fragmented Container Supplier Process Control Periodic Audit Of 3 10 3 90 Material Unpredictable Deployment Supplier Material Contaminated Fragmented Container Open Boxes Visual Inspection 1 9 7 63 Material Unpredictable Deployment Material Fragmented Container Engineering Change Release Verification 1 10 7 70 Composition Unpredictable Deployment Supplier Change Green "OK" Tag Change Customer Notification 2 Move To Unreleased Fragmentation Untrained LTO Check For Green "OK" 5 10 1 50 Approved Untrained Personnel Tag At Press Storage Trace Card Check List Training Leading to MSA. Critical features are determined by the FMEA (RPN indicators) and put into the Control Plan.
  • 47. Control Plan / Flow Diagram  Inspection Points  Inspection Frequency  Instrument  Measurement Scale  Sample Preparation  Inspection/Test Method  Inspector (who?)  Method of Analysis 47
  • 48. GM Process Flow Chart Process Flow Diagram Approved By: Part Number: Date: 4/5/93 QA Manager Part Description: Rev. : C Operations Manager Prepared By: Senior Advisor QA Engineer Step Fabrication Move Store Inspect Operation Description Item # Key Product Characteristic Item # Key Control Characteristic 1 Move "OK" Vinyl Material 1.0 Material Specs 1.0 Material Certification Tag From Storage Area and Load Into Press. 2 Auto Injection Mold Cover 2.0 Tearstrip In Cover 2.1 Tool Setup In Tool # 2.2 Machine Setup 3.0 Hole Diameter In Cover 2.1 Tool Setup 2.2 Machine Setup 4.0 Flange Thickness In Cover 2.1 Tool Setup 2.2 Machine Setup 5.0 Pressure Control Protrusions 2.1 Tool Setup Height 2.2 Machine Setup 3 Visually Inspect Cover 6.0 Pressure Control Protrusions 2.1 Tool Setup Filled Out 2.2 Machine Setup
  • 49. Standard Control Plan Example Control Plan Number Key Contact / Phone Date (Orig.) Date (Rev.) Part No./ Latest Change No. Core Team Customer Engineering Approval/Date Part Name/Description Supplier/Plant Apoproval/Date Customer Quality Approval/Date Supplier/Plant Supplier Code Other Approval/date (If Req'd) Other Approval/date (If Req'd) Characteristics Methods Part/ Process Number Process Name/ Operation Description Machine, Device, Jig, Tools for Mfg. No. Product Process Special Char. Class Product/ Process Spec/ Tolerance Evaluation Measurement Technique Size Frequ- ency Control Method Reaction Plan This form is on course disk 49
  • 51. Measurement as a System  Choosing the Right Instrument  Instrument Calibration Needs  Standards or Masters Needed  Accuracy and Precision  Measurement Practices  Where  How Many Places  Reported Figures  Significant Figures Rule  2 Action Figures  Rule of 10  Individuals, Averages, High-Lows 51
  • 52. Measurement Error Measured Value (y) = True Value (x) + Measurement Error Deming says there is no such thing as a ‘True’ Value. Consistent (linear)? 52
  • 53. Sources of Measurement Error  Sensitivity (Threshold) Chemical Indicators  Discrimination  Precision (Repeatability)  Accuracy (Bias)  Damage  Differences in use by Inspector (Reproducibility) Training Issues  Differences Among Instruments and Fixtures  Differences Among Methods of Use  Differences Due to Environment 53
  • 54. Types of Measurement Scales  Variables  Can be measured on a continuous scale  Defined, standard Units of Measurement  Attributes  No scale  Derived ‘Unit of Measurement’  Can be observed or counted  Either present or not  Needs large sample size because of low information content 54
  • 55. How We Get Data Inspection Measurement Test Includes Sensory (e.g..: look, touch, smell…etc) Magnitude of Quality 55
  • 56. Operational Definitions  Is the container Round?  Is your software Accurate?  Is the computer screen Clean?  Is the truck On Time? 56
  • 57. Different Method = Different Results In Spec Out of Spec Method 1 Method 2 57
  • 58. Measurement System Variability  Small with respect to Process Variation  Small with respect to Specified Requirements  Must be in Statistical Control Measurement IS a Process! Free of Assignable Causes of variation 58
  • 59. Studying the Measurement System  Environmental Factors  Human Factors  System Features  Measurement Studies 59
  • 60. Environmental Factors  Temperature  Humidity  Vibration  Lighting  Corrosion  Wear  Contaminants Oil & Grease Aerosols Where is the study performed? 1. Lab? 2. Where used? 3. Both? 60
  • 61. Human Factors  Training  Skills  Fatigue  Boredom  Eyesight  Comfort  Complexity of Part  Speed of Inspection (parts per hour)  Misunderstood Instructions 61
  • 62. Human Measurement Errors  Sources of Errors Inadvertent Errors  Attentiveness  Random  Good Mistake-Proofing Target Technique Errors  Consistent Wilful Errors (Bad mood)  Error Types (Can be machine or human) Type I - Alpha Errors [ risk] Type II - Beta Errors [ risk] Accept Reject Good Bad OK! OK! alpha beta Training Issue Process in control, but needs adjustment, False alarm Unaware of problem 62
  • 63. Measurement System Features  Discrimination Ability to tell things apart  Bias [per AIAG] (Accuracy)  Repeatability [per AIAG] (Precision)  Reproducibility  Linearity  Stability 63
  • 64. Discrimination  Readable Increments of Scale  If Unit of Measure is too course: Process variation will be lost in Rounding Off  The “Rule of Ten”: Ten possible values between limits is ideal Five Possible Values: Marginally useful Four or Less: Inadequate Discrimination 64
  • 66. Range Charts & Discrimination Indicates Poor Precision 66
  • 67. Bias and Repeatability Precise Imprecise Accurate Inaccurate Bias You can correct for Bias You can NOT correct for Imprecision 67
  • 68. Bias  Difference between average of measurements and an Agreed Upon standard value  Known as Accuracy  Cannot be evaluated without a Standard  Adds a Consistent “Bias Factor” to ALL measurements  Affects all measurements in the same way Standard Value Measurement Scale Bias 68
  • 69. Causes of Bias  Error in Master  Worn components  Instrument improperly calibrated  Instrument damaged  Instrument improperly used  Instrument read incorrectly  Part set incorrectly (wrong datum) 69
  • 70. Bias  Bias - The difference between the observed Average of measurements and the master Average of the same parts using precision instruments. (MSA Manual Glossary)  The auditor may want evidence that the concept of bias is understood. Remember that bias is basically an offset from ‘zero’. Bias is linked to Stability in the sense that an instrument may be ‘zeroed’ during calibration verification. Knowing this we deduce that the bias changes with instrument use. This is in part the concept of Drift. 70
  • 71. Bias  I choose a caliper (resolution 0.01) for the measurement. I measure a set of parts and derive the average.  I take the same parts and measure them with a micrometer (resolution 0.001). I then derive the average.  I compare the two averages. The difference is the Bias. 71
  • 72. Repeatability  Variation among repeated measurements  Known as Precision  Standard NOT required  May add or subtract from a given measurement  Affects each measurement randomly Measurement Scale Repeatability Margin of Error Doesn’t address Bias 5.15 = 99% 72
  • 73. Repeatability Issues  Measurement Steps  Sample preparation  Setting up the instrument  Locating on the part  How much of the measurement process should we repeat? 73
  • 74. Using Shewhart Charts I Repeatability 74
  • 76. Evaluating Bias & Repeatability  Same appraiser, Same part, Same instrument  Multiple readings (n≥10 with 20 to 40 better)  Analysis  Average minus Standard Value = Bias  5.15* Standard Deviation = Repeatability  or +/- 2.575  [99% repeatability]  or +/- 2  [95% repeatability]  Histogram  Probability AIAG 76
  • 77. Repeatability Issues  Making a measurement may involve numerous steps  Sample preparation  Setting up the instrument  Locating the part, etc.  How much of the measurement process should we repeat? How far do we go? 77
  • 78. Bias & Repeatability Histogram Never include assignable cause errors
  • 79. Linearity  The difference in the Bias or Repeatability across the expected operating range of the instrument. 79
  • 80. Plot Biases vs. Ref. Values Linearity = |Slope| * Process Variation = 0.1317*6.00 = 0.79 % Linearity = 100 * |Slope| = 13.17% 80
  • 81. Causes of Poor Linearity  Instrument not properly calibrated at both Upper and Lower extremes  Error in the minimum or maximum Master  Worn Instrument  Instrument design characteristics 81
  • 82. Reproducibility  Variation in the averages among different appraisers repeatedly measuring the same part characteristic  Concept can also apply to variation among different instruments Includes repeatability which must be accounted for. 82
  • 84. Calculating Reproducibility (I)  Find the range of the appraiser averages (R0)  Convert to Standard Deviation using d2* (m=# of appraisers; g=# of ranges used = 1)  Multiply by 5.15  Subtract the portion of this due to repeatability 84
  • 86. Stability  Variation in measurements of a single characteristic  On the same master  Over an extended period of time  Evaluate using Shewhart charts 86
  • 87. Evaluate Stability with Run Charts 87
  • 88. Stability Both gages are stable, but..... 88
  • 89. Importance of Stability  Statistical stability, combined with subject-matter knowledge, allows predictions of process performance  Action based on analysis of Unstable systems may increase Variation due to ‘Tampering’  A statistically unstable measurement system cannot provide reliable data on the process 89
  • 91. Analysis Tools  Calculations of Average and Standard Deviation  Correlation Charts  Multi-Vari Charts  Box-and-Whisker Plots  Run charts  Shewhart charts 91
  • 92. Average and Standard Deviation 92
  • 93. Correlation Charts  Describe Relationships  Substitute measurement for desired measurement  Actual measurement to reference value  Inexpensive gaging method versus Expensive gaging method  Appraiser A with appraiser B 93
  • 94. Substitute Measurements  Cannot directly measure quality  Correlate substitute measure  Measure substitute  Convert to desired quality 94
  • 95. Comparing Two Methods  Two methods  Measure parts using both  Correlate the two  Compare to “Line of No Bias”  Investigate differences Magnetic Stripping Line of Perfect Agreement Line of Correlation 95
  • 97. Measurements vs. Reference Correlation Disparity 97
  • 99. Run Charts Examine Stability 99
  • 100. Multiple Run Charts More than 3 appraisers confuses things...
  • 101. Multi-Vari Charts  Displays 3 points  Length of bar; bar-to-bar; Bar cluster to cluster  Plot High and Low readings as Length of bar  Each appraiser on a separate bar  Each piece in a separate bar cluster High Reading Low Reading Average Reading 101
  • 102. Multi-Vari Type I  Bar lengths are long  Appraiser differences small in comparison  Piece-to-piece hard to detect  Problem is repeatability 102
  • 103. Multi-Vari Type II  Appraiser differences are biggest source of variation  Bar length is small in comparison  Piece-to-piece hard to detect  Problem is reproducibility 103
  • 104. Multi-Vari Type III  Piece-to-piece variation is the biggest source of variation  Bar length (repeatability) is small in comparison  Appraiser differences (bar-to-bar) is small in comparison  Ideal Pattern 104
  • 106. Multi-Vari Chart, Joined Look for similar pattern 106
  • 107. Using Shewhart Charts  Subgroup = Repeated measurements,, same piece  Different Subgroups = Different pieces and/or appraisers  Range chart shows precision (repeatability)  Average chart “In Control” shows reproducibility If subgroups are different appraisers  Average chart shows discriminating power If subgroups are different pieces (“In Control” is BAD!) 107
  • 108. Shewhart Charts This is not a good way to plot this data Too many lines 108
  • 109. Shewhart Chart of Instrument 109
  • 111. Gauge R&R Studies  Developed by Jack Gantt  Originally plotted on probability paper  Revived as purely numerical calculations  Worksheets developed by AIAG  Renewed awareness of Measurement Systems as ‘Part of the Process’ Consider Numerical vs. Graphical Data Evaluations 111
  • 112. Terms Used in R&R (I)  n = Number of Parts [2 to 10] Parts represent total range of process variation Need not be “good” parts. Do NOT use consecutive pieces. Screen for size  a = Number of Appraisers Each appraiser measures each part r times Study must be by those actually using  R - Number of trials  Also called “m” in AIAG manual  g = r*a [Used to find d2* in table 2, p. 29 AIAG manual] 1 Outside Low/High 1 Inside Low/High Target Minimum of 5. 2 to 10 To accommodate worksheet factors 1 2 3 4 5 112
  • 113. Terms Used in R&R (II)  R-barA = Average range for appraiser A, etc.  R-double bar = Average of R-barA, R-barB  Rp = Range of part averages  XDIFF = Difference between High & Low appraiser averages Also a range, but “R” is not used to avoid confusion  EV = 5.15 = Equipment variation (repeatability)  EV = 5.15 = Equipment variation (reproducibility)  PV = Part variation  TV = Total variation Process Variation 113
  • 114. R&R Calculations Measurement System Variation Product Process Variation Left over Repeatability Remember - Nonconsecutive Pieces Left over Repeatability 114
  • 116. Evaluating R&R  %R&R=100*[R&R/TV] (Process Control)  %R&R=100*[R&R/Tolerance] (Inspection)  Under 10%: Measurement System Acceptable  10% to 30%: Possibly acceptable, depending upon use, cost, etc.  Over 30%: Needs serious improvement 116
  • 117. Analysis of Variance I  Mean squares and Sums of squares  Ratio of variances versus expected F-ratio  Advantages Any experimental layout Estimate interaction effects  Disadvantages Must use computer Non-intuitive interpretation 117
  • 118. Analysis of Variance II  The n*r measurements must be done in random sequence [a good idea anyway]  Assumes that EV [repeatability] is normal and that EV is not proportional to measurement [normally a fairly good assumption]  Details beyond scope of this course 118
  • 119. Special Gauging Situations  Go/No-Go  Destructive Testing 119
  • 120. If Gauges were Perfect 120
  • 121. But Repeatability Means We Never Know The Precise Value 121
  • 122. So - Actual Part Acceptance Will Look Like This: 122
  • 123. The Effect of Bias on Part Acceptance 123
  • 124. Go/No-Go gauges  Treat variables like attributes  Provide less information on the process, but...  Are fast and inexpensive  Cannot use for Process Control  Can be used for Sorting purposes 124
  • 125. “Short” Go/No-Go Study  Collect 20 parts covering the entire process range  Use two inspectors  Gage each part twice  Accept gauge if there is agreement on each of the 20 parts * May reject a good measuring system 125
  • 126. Destructive Tests  Cannot make true duplicate tests  Use interpenetrating samples  Compare 3 averages  Adjust using √n 126
  • 127. Destructive Tests: Interpreting Samples AIAG does not address 127
  • 129. Measurement Variation  Observed variation is a combination of the production process PLUS the measurement process  The contribution of the measurement system is often overlooked 129
  • 130. Types of Measurement Variation  Bias (Inaccuracy)  Repeatability (Imprecision)  Discrimination  Linearity  Stability 130
  • 131. Measurement Systems  Material  Characteristic  Sampling and Preparation  Operational Definition of Measurement  Instrument  Appraiser  Environment and Ergonomics 131
  • 132. Measurement Systems Evaluation Tools  Histograms  Probability paper  Run Charts  Scatter diagrams  Multi-Vari Charts  Gantt “R&R” analysis  Analysis of Variance (ANOVA)  Shewhart “Control” Charts 132
  • 133. Shewhart Charts  Range chart shows repeatability  X-bar limits show discriminating power  X-double bar shows bias (if a known standard exists)  Average chart shows stability (sub-groups overtime)  Average chart shows reproducibility (sub-groups over technicians/instruments) 133
  • 134. Conclusion  Rule of Ten  Operating Characteristic Curve  Special Problems Go/No-Go Gages Attribute Inspection Destructive Testing 134