Professional Documents
Culture Documents
Definitions
Definitions
History
History
Strategy
Strategy
Problem
ProblemSolving
Solving
Roles
Roles&&Responsibilities
Responsibilities
Selecting Projects
Elements of Waste
2
What is Six Sigma…as a Symbol?
Narrow
NarrowVariation
Variation Wide
WideVariation
Variation
The variation in a process refers to how tightly all the various outcomes are clustered around the average.
No process will produce the EXACT same output each time.
3
What is Six Sigma…as a Value?
Sigma can be used interchangeably with the statistical term Standard Deviation.
Standard Deviation is the average distance of data points away from the Mean in a
distribution.
Point of Inflection
4
What is Six Sigma…as a Measure?
-6 -5 -4 -3 -2 -1 +1 +2 +3 +4 +5 +6
The higher the sigma level, the better the performance. Six Sigma refers to a process having
6 Standard Deviations between the average of the process center and the closest specification
limit or service level.
5
Measure
6 Sigma
5 Sigma
4 Sigma
3 Sigma
2 Sigma
1 Sigma
The likelihood of a defect decreases as the number of Standard Deviations that can be fit
between the Mean and the nearest spec limit increases.
6
What is Six Sigma…as a Metric?
Each of these metrics serves a different purpose and may be used at different levels in the
organization to express the performance of a process in meeting the organization’s (or
customer’s) requirements. We will discuss each in detail as we go through the course.
• Defects
• Defects per unit (DPU) 20
18
• Parts per million (PPM)
16
• Defects per million opportunities (DPMO) 14
8
• Sigma (s) 0 20 40 60 80 100
These are certain metrics that we use in Six Sigma. You will learn more
about these through the course of your study.
7
What is Six Sigma…as a Benchmark?
Source: Journal for Quality and Participation, Strategy and Planning Analysis
8
What is Six Sigma…as a Method?
DMAIC provides the method for applying the Six Sigma philosophy
in order to improve processes.
9
What is Six Sigma…as a Tool?
- Six Sigma has not created new tools, it has simply organized a variety
of existing tools to create flow.
Customer Value
Management Product Process Process System Functional
Responsiveness,
Cost, Quality,
= EBIT, (Enabler) , Design , Yield , Speed , Uptime , Support
Delivery
10
What is Six Sigma…as a Goal?
Sweet Fruit
Design for Six Sigma
5+ Sigma
Bulk of Fruit
Process
3 - 5 Sigma Characterization
and Optimization
Ground Fruit
1 - 2 Sigma Simplify and
Standardize
11
What is Six Sigma…as a Philosophy?
General Electric: First, what it is not. It is not a secret society, a slogan or a cliché. Six Sigma is a
highly disciplined process that helps us focus on developing and delivering near-perfect products and
services. The central idea behind Six Sigma is that if you can measure how many "defects" you have in
a process, you can systematically figure out how to eliminate them and get as close to "zero defects" as
possible. Six Sigma has changed the DNA of GE — it is now the way we work — in everything we do
and in every product we design.
Honeywell: Six Sigma refers to our overall strategy to improve growth and productivity as well as a
measurement of quality. As a strategy, Six Sigma is a way for us to achieve performance breakthroughs. It
applies to every function in our company, not just those on the factory floor. That means Marketing,
Finance, Product Development, Business Services, Engineering and all the other functions in our
businesses are included.
Lockheed Martin: We’ve just begun to scratch the surface with the cost-saving initiative called Six
Sigma and already we’ve generated $64 million in savings with just the first 40 projects. Six Sigma uses
data gathering and statistical analysis to pinpoint sources of error in the organization or products and
determines precise ways to reduce the error.
12
History of Six Sigma
• 1984 Bob Galvin of Motorola edicted the first objectives of Six Sigma
– 10x levels of improvement in service and quality by 1989
– 100x improvement by 1991
– Six Sigma capability by 1992
– Bill Smith, an engineer from Motorola, is the person credited as the father of Six Sigma
• 1984 Texas Instruments and ABB Work closely with Motorola to further develop
Six Sigma
• 1994 Application experts leave Motorola
• 1995 AlliedSignal begins Six Sigma initiative as directed by Larry Bossidy
– Captured the interest of Wall Street
• 1995 General Electric, led by Jack Welsh, began the most widespread
undertaking of Six Sigma even attempted
• 1997 To present Six Sigma spans industries worldwide
13
History of Six Sigma
14
History of Six Sigma
• Six Sigma created a realistic and quantifiable goal in terms of its target of 3.4 defects
per million operations. It was also accompanied by a methodology to attain that
goal.
• That methodology was a problem solving strategy made up of four steps: measure,
analyze, improve and control.
• When GE launched Six Sigma they improved the methodology to include the Define
Phase.
15
DMAIC Phases Roadmap
Champion/
Process
Owner
Estimate COPQ
Charter Project
Measure
16
Define Phase Deployment
Business Case
Selected
N Estimate COPQ
Approved
Project Recommend Project Focus
Focus
Y
Create Team
Charter Team
17
Define Phase Deliverables
Listed below are the type of Define Phase deliverables that will be
reviewed by this course.
By the end of this course, you should understand what would be necessary
to provide these deliverables in a presentation.
– Charter Benefits Analysis
– Team Members (Team Meeting Attendance)
– Process Map – high level
– Primary Metric
– Secondary Metric(s)
– Lean Opportunities
– Stakeholder Analysis
– Project Plan
– Issues and Barriers
18
Six Sigma Strategy
Requirement Requirement
or or
LSL
Target USL
20
Problem Solving Strategy
Y=f (Xi)
This simply states that Y is a function of the X’s. In
other words Y is dictated by the X’s.
21
Example
Y=f (Xi)
Which process variables (causes) have critical impact on the output
(effect)?
Crusher Yield
= f ( Feed , Speed , Material
Type ,
Tool
Wear , Lubricant )
Time to Close =f( Trial
,
Correct Sub
,
Credit
,
Entry
Balance Accounts Accounts Memos Mistakes Xn, , )
Applied
Espresso =f( X1 , X2 , X3 , X4 , Xn )
23
Six Sigma Strategy
(X6)
(X2)
24
Breakthrough Strategy
Bad 6-Sigma
6-Sigma
Breakthrough UCL
UCL
Breakthrough
Old Standard
Performance
LCL
UCL
UCL
New Standard
LCL
LCL
Good
25
VOC, VOB, VOE
The foundation of Six Sigma requires Focus on the voices of the Customer, the Business and the Employee which provides:
26
Six Sigma Roles and Responsibilities
• Executive Leadership
MBB
• Champion/Process Owner
Black Belts • Master Black Belt
• Black Belt
Green Belts
• Green Belt
Yellow Belts • Yellow Belt
Not all Six Sigma deployments are driven from the top by Executive
Leadership. The data is clear, however, that those deployments that are
driven by executive management are much more successful than those that
aren’t.
• Makes decision to implement the Six Sigma initiative and develop
accountability method
• Eliminates barriers
28
Champion/Process Owner
Champions identify and select the most meaningful projects to work on,
they provide guidance to the Six Sigma belt and open the doors for the
belts to apply the process improvement technologies.
29
Master Black Belt
MBB should be well versed with all aspects of Six Sigma, from technical
applications to Project Management. MBBs need to have the ability to
influence change and motivate others.
30
Black Belt
Black Belts are application experts and work projects within the business.
They should be well versed with The Six Sigma Technologies and have the
ability to drive results.
• Works cross-functionally
31
Green Belt
32
Yellow Belt
33
The Life of a Six Sigma Belt
• Must complete two projects within one year from beginning of training
35
Organizational Behaviors
All players in the Six Sigma process must be willing to step up and act
according to the Six Sigma set of behaviors.
36
Summary
37
Six Sigma Fundamentals
Process Maps
Process Metrics
Selecting Projects
Elements of Waste
38
What is a Process?
39
Examples of Processes
40
Process Maps
• Process Maps are living documents and must be changed as the process is
changed
– They represent what is currently happening, not what you think is happening.
– They should be created by the people who are closest to the process
Process Map
t
Start Step A Step B Step C Step D Finish
ec
sp
In
41
Process Map Symbols
42
High Level Process Map
One of the deliverables from the Define Phase is a high level Process Map, which
at a minimum must include:
– Start and stop points
– All process steps
– All decision points
– Directional flow
– Value categories as defined below
• Value Added:
– Physically transforms the “thing” going through the process
– Must be done right the first time
– Meaningful from the customer’s perspective (is the customer willing to pay for it?)
• Value Enabling:
– Satisfies requirements of non-paying external stakeholders (government regulations)
• Non-Value Added
– Everything else
43
Process Map Example
N WALK-IN NOTE
CALL or DATA ENDS ENTER APPROPRIATE
WALK-IN? N SSAN (#,9s,0s)
Z CALL PUT ON HOLD,
REFER TO IF EMP DATA NOT
PHONE DATA REFERENCES POPULATED, ENTER
CAPTURE BEGINS
CREATE A CASE
Y INCL CASE TYPE
ANSWER? OLD N
DETERMINE WHO DATE/TIME, &
CASE
IS INQUIRING N NEEDED BY
Y
QUERY INTERNAL UPDATE ENTRIES
ACCESS CASE TOOL HRSC SME(S) INCL OPEN DATE/TIME AUTO Y
ROUTE
ROUTE
DETERMINE NATURE N
OF CALL & CONFIRM Y
ANSWER?
UNDERSTANDING
CASE Y CLOSE CASE
N CLOSED W/ E
DATE/TIME
CASE TOOL N OFF HOLD AND ADD TO N
RECORD? C ARRANGE CALL RESEARCH
BACK PHONE DATA LIST GO TO E
TAKE ACTION
Y ENDS F or E NEXT
or
DEPENDING ON
DO RESEARCH F
B CASE
44
Cross Functional Process Map
When multiple departments or functional groups are involved in a complex process it is often
useful to use cross functional Process Maps.
– Draw in either vertical or horizontal Swim Lanes and label the functional groups and draw the
Process Map
Produce an
Invoice
No
enrollment payment End
form
Match against
Accounting
Maintain database
Financial
Accepts transactions,
Bank
Review and
General
21.0
Process 3.0
Bank
transfer in Journey Entry
Reconciliation
FRS
45
Process Map Exercise
46
Do you know your Customer?
47
Voice of the Customer
Voice of the Customer or VOC seems obvious; after all, we all know what the
customer wants. Or do we??
The customer’s perspective has to be foremost in the mind of the Six Sigma Belt
throughout the project cycle.
1. Features
• Does the process provide what the customers expect and need?
• How do you know?
2. Integrity
• Is the relationship with the customer centered on trust?
• How do you know?
3. Delivery
• Does the process meet the customer’s time frame?
• How do you know?
4. Expense
• Does the customer perceive value for cost?
• How do you know?
48
What is a Customer?
There are different types of customers which dictates how we interact with them in
the process. In order to identify customer and supplier requirements we must first
define who the customers are:
External
– Direct: those who receive the output of your services, they generally are the
source of your revenue
– Indirect: those who do not receive or pay for the output of your services but
have a vested interest in what you do (government agencies)
Internal
- those within your organization
who receive the output of your
work
49
Value Chain
The relationship from one process to the next in an organization creates a “Value Chain” of
suppliers and receivers of process outputs.
Each process has a contribution and accountability to the next to satisfy the external customer.
External customers needs and requirements are best met when all process owners work
cooperatively in the Value Chain.
Careful –
each move
has many
impacts!
50
What is a CTQ?
• Critical to Quality (CTQ ’s) are measures that we use to capture VOC properly.
(also referred to in some literature as CTC’s – Critical to Customer)
• CTQ ’s can be vague and difficult to define.
– The customer may identify a requirement that is difficult to measure directly so it will
be necessary to break down what is meant by the customer into identifiable and
measurable terms
Product: Service:
• Performance • Competence
• Features • Reliability
• Conformance • Accuracy
• Timeliness • Timeliness
• Reliability • Responsiveness
• Serviceability • Access
• Durability • Courtesy
• Aesthetics • Communication
• Reputation • Credibility
• Completeness • Security
• Understanding
51
Developing CTQ’s
Identify Customers
Step 1 • Listing
• Segmentation
• Prioritization
Validate CTQ s
Step 2 • Translate VOC to CTQ’s
• Prioritize the CTQ’s
• Set Specified Requirements
• Confirm CTQ’s with customer
Capture VOC
Step 3 • Review existing performance
• Determine gaps in what you need to know
• Select tools that provide data on gaps
• Collect data on the gaps
52
Cost of Poor Quality (COPQ)
• As a Six Sigma Belt, one of your tasks will be to estimate COPQ for your process
• Through your process exploration and project definition work you will develop a
refined estimate of the COPQ in your project
• Calculating COPQ is iterative and will change as you learn more about the
process
• The concepts of traditional Quality Cost are the foundation for COPQ.
– External, Internal, Prevention, Appraisal
• A significant portion of COPQ from any defect comes from effects that are
difficult to quantify and must be estimated.
54
COPQ - Categories
55
COPQ - Iceberg
Inspection
Warranty Recode
Rework
Rejects
Visible Costs
More Set-ups
Excess inventory
Excessive Material
Orders/Planning
Hidden Costs Lost Customer Loyalty
56
COPQ and Lean
Waste does not add, subtract or otherwise modify the throughput in a way that is
perceived by the customer to add value.
57
COPQ – Hard and Soft Savings
While hard savings are always more desirable because they are easier to
quantify, it is also necessary to think about soft savings.
58
COPQ Exercise
59
The Basic Six Sigma Metrics
• Better: DPU, DPMO, RTY (there are others, but they derive from these basic three)
• Faster: Cycle Time
• Cheaper: COPQ
IfIfyou
youmake
makethe
theprocess
processbetter
betterby
byeliminating
eliminatingdefects
defectsyou
youwill
willmake
makeititfaster.
faster.
IfIfyou
youchoose
choosetotomake
makethe
theprocess
processfaster,
faster,you
youwill
willhave
havetotoeliminate
eliminatedefects
defectstotobe
beasasfast
fastas
as
you
youcancanbe.
be.
IfIfyou
youmake
makethe
theprocess
processbetter
betteror
orfaster,
faster,you
youwill
willnecessarily
necessarilymake
makeititcheaper.
cheaper.
The
Themetrics
metricsfor
forall
allSix
SixSigma
Sigmaprojects
projectsfall
fallinto
intoone
oneof
ofthese
thesethree
three
categories.
categories.
60
Cycle Time Defined
– It is the time required for the product or transaction to go through the entire process,
from beginning to end
– It is not simply the “touch time” of the value-added portion of the process
61
Defects Per Unit (DPU)
Six Sigma methods quantify individual defects and not just defectives
– Defects account for all errors on a unit
• A unit may have multiple defects
• An incorrect invoice may have the wrong amount due and the wrong due date
– Defectives simply classifies the unit bad
• Doesn’t matter how many defects there are
• The invoice is wrong, causes are unknown
– A unit:
• Is the measure of volume of output from your area.
• Is observable and countable. It has a discrete start and stop point.
• It is an individual measurement and not an average of measurements.
62
First Time Yield
*None of the data used herein is associated with the products shown herein. Pictures are no more than illustration to make a point to teach the concept.
63
Rolled Throughput Yield
RTY = X1 * X2 * X3
Units in = 100 Units in = 100 Units in = 100
Units W/O Rework = 60 Units W/O Rework = 70 Units W/O Rework = 80 Units Passed = 34
RTY = 0.6 RTY = 0.7 RTY = 0.8 Units Tested = 100
Process A (Grips) Process B (Shafts) Process C (Club Heads) Final Product (Set of Irons)
*None of the data used herein is associated with the products shown herein. Pictures are no more than illustration to make a point to teach the concept.
64
RTY Estimate
RTY = e-dpu
65
Deriving RTY from DPU
The Binomial distribution is the true model for defect data, but the Poisson is the convenient
model for defect data. The Poisson does a good job of predicting when the defect rates are low.
120%
Poisson
Poisson VS
VS Binomial
Binomial (r=0,n=1)
(r=0,n=1) Probability
Probability Yield
Yield Yield
Yield %
% Over
Over
120% of
of aadefect
defect (Binomial)
(Binomial) (Poisson)
(Poisson) Estimated
Estimated
0.0
0.0 100%
100% 100%
100% 0%
0%
100%
100% Yield
Yield (Binomial)
(Binomial) 0.1
0.1 90%
90% 90%
90% 0%
0%
Yield
Yield (Poisson) 0.2 80% 82% 2%
(RTY)
80%
80%
0.3
0.3 70%
70% 74%
74% 4%
4%
60% 0.4
0.4 60%
60% 67%
67% 7%
7%
60%
0.5 50% 61% 11%
Yield
Binomial
n = number of units
r = number of predicted defects
p = probability of a defect occurrence Poisson
q=1-p
For low defect rates (p < 0.1), the Poisson approximates the Binomial fairly well.
66
Deriving RTY from DPU - Modeling
Unit
Basic Question: What is the likelihood of producing a
Opportunity unit with zero defects?
Yield
0.356
0.352
• What is probability that any given opportunity 0.348
will be a defect? 10
10 100
100 1000
1000 10000
10000 100000
100000 1000000
1000000
Chances Per Unit
68
Six Sigma Metrics – Calculating DPU
The DPU for a given operation can be calculated by dividing the number of defects found in
the operation by the number of units entering the operational step.
Product A
FTY = 80%
Product B
FTY = 80%
*None of the data used herein is associated with the products shown herein. Pictures are no more than illustration to make a point to teach the concept.
70
Focusing our Effort – FTY vs. RTY
Let’s look at the DPU of each product assuming equal opportunities and margin…
Product A
Product B
*None of the data used herein is associated with the products shown herein. Pictures are no more than illustration to make a point to teach the concept.
71
Summary
• Explain COPQ
72
Project Selection
Selecting Projects
Selecting Projects
Financial Evaluation
Elements of Waste
73
Approaches to Project Selection
“Structured Approach”
Identifies projects based on organizational data, provides a direct plan to effect core
business metrics that have bottom-line impact.
Business Case – The Business Case is a high level articulation of the area of
concern. This case answers two primary questions; one, what is the business
motivation for considering the project and two, what is our general area of
focus for the improvement effort?
Project Charter – The Project Charter is a more detailed version of the Business
Case. This document further focuses the improvement effort. It can be
characterized by two primary sections, one, basic project information and, two,
simple project performance metrics.
75
Project Selection - Governance
Responsible Frequency
Party Resources of Update
The Starting Point is defined by the Champion or Process Owner and the Business Case is the
output.
– These are some examples of business metrics or Key Performance Indicators commonly referred
to as KPI’s.
– The tree diagram is used to facilitate the process of breaking down the metric of interest.
EBIT
Level 2
Cycle time
Defects Level 2
Cost Level 1
Level 2
Revenue
Complaints Level 2
Compliance
Safety
77
A Structured Approach - Snapshot
Business Measures
Actionable Level
Key Performance Indicators (KPIs)
78
Business Case Components – Level 1
– Primary business measure should bear a direct line of site with the organizations strategic
objective.
– As the Champion narrows in on the greatest opportunity for improvement, this provides a
clear focus for how the success will be measured.
79
Business Case Components – Business Measures
Business Business
Activities Processes
Primary Business Measure Measure
80
Business Case Components - Activities
Business Business
Activities Processes
Primary Business Measure Measure
81
Business Case Components - Processes
Business Business
Activities Processes
Primary Business Measure Measure
82
What is a Business Case?
The Business Case communicates the need for the project in terms of
meeting business objectives.
83
Business Case Example
84
The Business Case Template
85
Business Case Exercise
1. Complete the Business Case template below to the best of your ability.
86
What is a Project Charter?
87
Project Charter - Definitions
• Problem Statement - Articulates the pain of the defect or error in the process.
• Objective Statement – States how much of an improvement is desired from the project.
• Primary Metric – The actual measure of the defect or error in the process.
• Charts – Graphical displays of the Primary and Secondary Metrics over a period of time.
88
Project Charter - Problem Statement
89
Project Charter – Objective & Scope
Pareto Analysis:
• A bar graph used to arrange information in such a way that priorities for process
improvement can be established.
Pareto Chart of Help Text Errors
400
100
300 80
Quantity
Percent
200 60
40
100
20
0 0
Help Text Errors
C2 145 142 34 29 8
Percent 40.5 39.7 9.5 8.1 2.2
Cum % 40.5 80.2 89.7 97.8 100.0
• The 80-20 theory was first developed in 1906, by Italian economist, Vilfredo Pareto, who
observed an unequal distribution of wealth and power in a relatively small proportion of
the total population. Joseph M. Juran is credited with adapting Pareto's economic
observations to business applications.
91
The 80:20 Rule Examples
92
Pareto Chart - Tool
Multi level Pareto Charts are used in a drill down fashion to get to Root
Cause of the tallest bar.
250000 80
200000
60
Percent
Scrap
150000
Pareto Chart
40 by Department
100000 Level 2
180000
160000 20 100
50000
140000
0 0 Pareto
80 Chart by Part Number
Count A B 120000 C Level 3
100000
Scrap 250000 30000 25000
Percent
100000 100
Count
40000
60000 20 60
Percent
20000
0 40000 0
Department J M F W Other 40
Count 95000 23000 19000 17500 5000
Percent 59.6 14.4 11.9 11.0 3.1
Cum % 59.6 74.0 85.9 20000 96.9 100.0 20
0 0
Part Z101 Z876 X492
Percent 78.9 15.8 5.3
Cum % 78.9 94.7 100.0
93
Pareto Chart - Tool
Percent
Scrap
160000 100
150000
140000
40
80 140000
100000 120000
80
120000
Percent
100000
Count
20 60
50000
80000 Pareto Chart by Part Number
Percent
Level 3 100000
Count
0 0 40
Count A B
60000
C
100000
100
60
Scrap 250000 30000 40000 25000
Percent 82.0 9.8
20000
8.2
80000
20 80000
Cum % 82.0 91.8 100.0 80
0 0
60000 40
Department J M F W Other
Count 95000 23000 60000
19000 17500 5000 60
Percent
Percent 59.6 14.4 11.9 11.0 3.1
Cum % 59.6 74.0 85.9 96.9 100.0 40000
40000 40 20
20000
20000 20
0 0
0 0
Department J M F W Other
Part
Percent
Z101
78.9
Z876
15.8
X492
5.3
Count 95000 23000 19000 17500 5000
• Level 1:
Pareto Chart by Part Number
Level 3
– Department J Makes up 60% of the 100000
100
Scrap 80000
80
• Level 2: 60000
60
Percent
– Part Z101 Makes up 80% of 40000 40
0 0
Part Z101 Z876 X492
Percent 78.9 15.8 5.3
Cum % 78.9 94.7 100.0
94
Pareto Chart - Example
95
Pareto Chart - Example
Percent
Count
60
1500
40
1000
500 20
0 0
FAILURE MODE
When your Pareto shows up like this your focus is probably too broad.
96
Pareto Chart - Example
– Using a higher level scope for the first Pareto may help in providing
focus.
– Create another Pareto as shown above.
97
Pareto Chart - Example
This gives a better picture of which product category produces the highest defect
count.
1500
Percent
60
Count
1000 40
500 20
0 0
PRODUCT CATAGORIES
Now we’ve got something to work with. Notice the 80% area, draw a line from
the 80% mark across to the cumulative percent line (Red Line) in the graph as
shown here.
98
Pareto Chart - Example
Now that we have more of a focus area, drill down one more level.
– This chart will only use the classifications within the first bar on the previous chart.
– Create another Pareto which will drill down to the categories within the Card type
from the previous Pareto.
99
Pareto Chart - Example
Pareto
Pareto Chart
Chart of
of TRAVEL
TRAVEL
1400
1400
100
100
1200
1200
1000
1000 80
80
Percent
800
Percent
800
Count
Count
60
60
600
600
40
40
400
400
20
20
200
200
00 00
TRAVEL
TRAVEL CAR
CAR HOTEL
HOTEL AIR
AIR
Count
Count 428
428 420
420 390
390
Percent
Percent 34.6
34.6 33.9
33.9 31.5
31.5
Cum
Cum %% 34.6
34.6 68.5
68.5 100.0
100.0
101
Project Charter – Secondary Metrics
102
Project Charter – Metric Charts
Generating Charts:
104
What is the Financial Evaluation?
105
Benefits Capture - Calculation “Template”
Whatever your organization’s protocol may be these aspects should be accounted for
within any improvement project.
106
Benefits Capture – Basic Guidelines
• The Project Measure (Primary Metric) has to have a direct link between
the process and its KPI’s.
107
Benefits Capture - Categorization
A
• Projects directly impact the Income Statement or Cash Flow Statement.
109
Benefits Capture - Summary
• Performance tracking for Six Sigma Projects should use the same
discipline that would be used for tracking any other high-profile
projects.
It’s a wrap!
110
Benefits Calculation Template
The Benefits Calculation Template facilitates and aligns with the aspects
discussed for Project Accounting.
111
Summary
• Refine and Define the business problem into a Project Charter to display
critical aspects of an improvement project
112
Elements of Waste
Selecting Projects
Elements of Waste
7 Components of Waste
5S
113
Definition of Lean
114
Lean – History
115
Lean Six Sigma
118
Overproduction
Overproduction is producing more than the next step needs or more than
the customer buys.
– It may be the worst form of waste because it contributes to all the others.
Examples are:
Over-ordering materials
Duplication of effort/reports
119
Correction
Misspelled words in
communications
Eliminate erors!!
Waste of Correction includes the waste of handling and fixing
mistakes. This is common in both manufacturing and
transactional settings.
120
Inventory
Inventory is the liability of materials that are bought, invested in and not
immediately sold or used.
Examples are:
Over-ordering materials
consumed in-house
Examples are:
Extra steps
122
Overprocessing
Examples are:
Sign-offs
Communications, reports,
emails, contracts, etc that contain
more than the necessary points
(briefer is better)
Waste of Overprocessing relates to over-processing
anything that may not be adding value in the eyes of Voice mails that are too long
the customer.
123
Conveyance
Examples are:
Distance traveled
124
Waiting
Examples are:
– Overproduction ___________________
– Correction ___________________
– Inventory ___________________
– Motion ___________________
– Overprocessing ___________________
– Conveyance ___________________
– Waiting ___________________
126
5S – The Basics
127
English Translation
There have been many attempts to force five English “S” words to maintain the original
intent of 5S from Japanese. Listed below are typical English words used to translate:
1.) Sort (Seiri)
2.) Straighten or Systematically Arrange (Seiton)
3.) Shine or Spic and Span (Seiso)
4.) Standardize (Seiketsu) Place things in such a way that
5.) Sustain or Self-Discipline (Shitsuke) they can be easily reached
whenever they are needed.
Straighten
Sort Shine
5S
Visual sweep of areas, eliminate
Identify necessary items and remove dirt, dust and scrap. Make
unnecessary ones, use time management. workplace shine.
Self-Discipline
Standardize
Make 5S strong in habit.
Work to standards, maintain
Make problems appear and
standards, wear safety
solve them.
equipment.
128
Exercise
• Sort ____________________
• Straighten ____________________
• Shine ____________________
• Standardize ____________________
• Self-Discipline ____________________
129
Summary
• Describe 5S
130
Define Phase
Wrap Up and Action Items
Define Phase Overview—The Goal
• At the completion of the Define Phase you should have a description of the
process defect that is creating waste for the business.
132
Define Action Items
At this point you should all understand what is necessary to complete these
action items associated with Define.
Look for the potential roadblocks and plan to address them before they become
problems:
– No historical data exists to support the project.
– Team members do not have the time to collect data.
– Data presented is the best guess by functional managers.
– Data is communicated from poor systems.
– The project is scoped too broadly.
– The team creates the “ideal” Process Map rather than the “as is” Process
Map.
Process Owner
Champion/
Estimate COPQ
Establish Team
Measure
136
Define Deployment
Business Case
Selected
N Estimate COPQ
Approved
Project Recommend Project Focus
Focus
Y
Create Team
Charter Team
137
Action Items Support List
138
Summary
139
Measure Phase
Process Discovery
Welcome to Measure
Process Discovery
FMEA
Process Capability
141
Overview of Brainstorming Techniques
A commonly used tool to solicit ideas by using categories to stimulate cause and
effect relationships within a problem. It uses verbal inputs in a team environment.
The Y
The or
Problem
The X’s Problem
Condition
(Causes)
l
Material Measurement Environment Categories
142
Cause and Effect Diagram
The Y
Theor
Problem
(Causes)
The Measurement category groups Root Causes related to the measurement and measuring of a
process activity or output:
Examples of questions to ask:
• Is there a metric issue? Measurement
• Is there a valid measurement
system? Is the data good enough?
• Is data readily available?
Y
The People category groups Root Causes related to people, staffing and
organizations:
Examples of questions to ask: People
• Are people trained, do they
have the right skills?
• Is there person to person
Y
variation?
• Are people over-worked?
144
Cause and Effect Diagram
The Method category groups Root Causes related to how the work is done, the way the process
is actually conducted:
Examples of questions to ask: Method
• How is this performed?
• Are procedures correct?
• What might unusual? Y
The Materials category groups Root Causes related to parts, supplies, forms or information
needed to execute a process:
Materials
145
Cause and Effect Diagram
The Equipment category groups Root Causes related to tools used in the process:
The Environment (a.k.a. Mother Nature) category groups Root Causes related to our work
environment, market conditions and regulatory issues.
Examples of questions to ask:
• Is the workplace safe and comfortable?
• Are outside regulations impacting the business? Y
• Does the company culture aid the process?
Environment
146
Classifying the X’s
The Cause & Effect Diagram is simply a tool to generate opinions about possible
causes for defects.
For each of the X’s identified in the Fishbone diagram classify them as follows:
– Controllable – C (Knowledge)
– Procedural – P (People, Systems)
– Noise – N (External or Uncontrollable)
Measurement
Capability (C) Adherence to procedure (P) Specifications (C)
Chemical
Startup inspection (P) Room Humidity (N) Column Capability (C) Purity
Handling (P) RM Supply in Market (N) Nozzle type (C)
Purification Method (P) Shipping Methods (C) Temp controller (C)
148
Cause & Effect Diagram - MINITAB™
Below is a Cause & Effect Diagram for surface flaws. The next few slides will
demonstrate how to create it in MINITAB™.
Fishbone Diagram
Measurem Material Personnel
ents
S hifts
A lloy s
S uperv isors
M icrometers
Lubricants
M icroscopes
Training
S uppliers
Inspectors O perators
Surface
Flaws
B rake
C ondensation S peed
E ngager
Lathes
M oisture%
B its
A ngle
S ockets
149
Cause & Effect Diagram - MINITAB™
Open the MINITAB™ Project “Measure Data Sets.mpj” and select the worksheet “Surfaceflaws.mtw”.
150
Cause & Effect Diagram - MINITAB™
151
Cause & Effect Diagram - MINITAB™
In order to adjust the Fishbone Diagram so the main causes titles are not rolled grab
the line with your mouse and move the entire bone.
Fishbone Diagram
Measurem Material Personnel
ents
S hifts
A lloy s
S uperv isors
M icrometers
Lubricants
M icroscopes
T raining
S uppliers
Inspectors O perators
Surface
F laws
B rake
C ondensation S peed
E ngager
Lathes
M oisture%
B its
A ngle
S ockets
152
Cause & Effect Diagram Exercise
1. Retrieve the high level Process Map for your project and use it to
complete a Fishbone, if possible include your project team.
153
Overview of Process Mapping
t
Step D Finish
ec
Start Step A Step B Step C
sp
In
154
Information from Process Mapping
1 2 3
What you THINK it is.. What it ACTUALLY is.. What it SHOULD be..
156
Standard Process Mapping Symbols
157
Process Mapping Levels
Level 2 – The Process Map, sometimes called the Worker level or viewpoint. This
example is from the perspective of the pizza chef
Pizza
Dough
No
Take Order Add Place in Observe Check Yes Remove
from Cashier Ingredients Oven Frequently if Done from Oven 1
Start New
Pizza
Scrap
No
Tape
Pizza Place in Put on
1 Correct Box
Order on Delivery Rack
Yes Box
Level 3 – The Micro Process Map, sometimes called the Improvement level or
viewpoint. Similar to a level 2, it will show more steps and tasks and on it will be
various performance data; yields, cycle time, value and non value added time, defects,
etc.
158
Types of Process Maps
As the name states, this diagram shows the process steps in a sequential flow, generally ordered from an upper left
corner of the map towards the right side.
Take
Order
Cook
Deliver
Pizza
The value of the Swim Lane map is that is shows you who or which department is responsible for the steps in a
process. This can provide powerful insights in the way a process performs. A timeline can be added to show how long
it takes each group to perform their work. Also each time work moves across a swim lane, there is a “Supplier –
Customer” interaction. This is usually where bottlenecks and queues form.
159
Process Maps – Examples for Different Processes
Scrap
Scrap
S u p p lPi ePr ror oc cu ur er emT moepen nMt tgF t i/n a n c e I . T . B Bu us isni ne se ss s
Prepare
paperwork Review &
U n it
Review &
Configure
approve
Review & & install
Configure
standard
approve
& install
standard
Review &
Issue
C o rp o ra te
approve
Review & payment
Issue
CAAR
C o rp o ra te
approve
payment
CAAR
Review &
approve
Review &
CAAR
approve
CAAR
Acquire
equipment
Acquire
equipment
S u p p lie r
Supplier Supplier
Ships
Supplier Paid
Supplier
Ships Paid
160
Types of Process Maps
r ATT Phones r Pizza type r See Below r Price r Cook r Complete call < 3 min
r Office Depot r Size r Order confirmation r Accounting r Order to Cook < 1 minute
r TI Calculators r Quantity r Bake order r Complete bake order
r NEC Cash Register r Extra Toppings r Data on cycle time r Correct bake order
r Special orders r Order rate data r Correct address
r Drink types & quantities r Order transaction r Correct Price
r Other products r Delivery info
r Phone number
r Address
r Name
r Time, day and date
r Volume
The SIPOC diagram is especially useful after you have been able to construct either a Level 1
or Level 2 Map because it facilitates your gathering of other pertinent data that is affecting the
process in a systematic way.
161
Types of Process Maps
The Value Stream Map
Process Steps
Log Route Disposition Cut Check Mail Delivery
-Computer -Department -Guidelines -Computer -Envelops
-1 Person Assignments -1 Person -Printer I -Postage
Size of work queue or I I -1 Person
I I -1 Person -1 Person
inventory
4,300 7,000 C/T = 75 sec 1,700 C/T = 255 sec 2,450 C/T = 15 sec 1,840 C/T = 100 sec
C/T = 15 sec
Uptime = 0.90 Uptime = 0.95 Uptime = 0.95 Uptime = 0.85 Uptime = 0.90
Process Step Time Hours = 8 Hours = 8 Hours = 8 Hours = 8 Hours = 8
Parameters Breaks = 0.5 Breaks = 0.5 Breaks = 0.5 Breaks = 0.5 Breaks = 0.5
Hours Hours Hours Hours Hours
Available =6.75 Available =7.13 Available =7.13 Available =6.38 Available =6.75
Sec. Sec. Sec. Sec. Sec.
Avail. = 24,300 Avail. = 25,650 Avail. = 25,650 Avail. = 22,950 Avail. = 24,300
Step Processing Time
15 sec 75 sec 255 sec 15 sec 100 sec
Days of Work in
queue 2.65 days 20.47 days 16.9 days 1.60 days 7.57 days
Cum Material Yield = .96 X .94 X .69 X .96 X .96 = .57 RTY = .92 X .94 X .59 X .96 X .96 = .47
The Value Stream Map is a very powerful technique to understand the velocity of process
transactions, queue levels and value added ratios in both manufacturing and non-
manufacturing processes.
162
Process Mapping Exercise – Going to Work
The purpose of this exercise is to develop a Level 1 Macro, Linear Process Flow Map and
then convert this map to a Swim Lane Map.
Read the following background for the exercise: You have been concerned about your
ability to arrive at work on time and also the amount of time it takes from the time your alarm
goes off until you arrive at work. To help you better understand both the variation in arrival
times and the total time, you decide to create a Level 1 Macro Process Map. For purposes of
this exercise, the start is when your alarm goes off the first time and the end is when you arrive
at your work station.
Task 1 – Mentally think about the various tasks and activities that you routinely do from the
defined start to the end points of the exercise.
Task 2 – Using a pencil and paper create a Linear Process Map at the macro level, but with
enough detail that you can see all the major steps of your process.
Task 3 – From the Linear Process Map, create a swim lane style Process Map. For the lanes
you may use the different phases of your process, such as the wake up phase, getting prepared,
driving, etc.
163
A Process Map of Process Mapping
Select the process Create the Level 2 PFM Create a Level 3 PFM
Determine approach to
map the process Perform SIPOC Add Performance data
Identify customer
Create Level 1 PFM
requirements
164
Process Mapping Approach
Select the
Using the Individual Approach
process 1. Start with the Level 1 Macro Process Map.
2. Meet with process owner(s) / manager(s). Create a Level 1
Map and obtain approval to interview process members.
Determine
approach to 3. Starting with the beginning of the process, pretend you are the
map the product or service flowing through the process, interview to
process
gather information.
4. As the interview progress, assemble the data into a Level 2
Complete PFM.
Level 1 PFM
worksheet 5. Verify the accuracy of the Level 2 PFM with the individuals
who provided input.
6. Update the Level 2 PFM as needed.
Create Level 1
PFM
165
Process Mapping Approach
Select the
Using the Team Approach
process 1. Start with the Level 1 Macro Process Map.
2. Meet with process owner(s) / manager(s). Create a Level 1
Map and obtain approval to call a process mapping meeting
Determine
approach to with process members (See team workshop instructions for
map the details on running the meeting).
process
3. Bring key members of the process into the process flow
workshop. If the process is large in scope, hold individual
Complete workshops for each subsection of the total process. Start with
Level 1 PFM the beginning steps. Organize meeting to use the “post-it note
worksheet
approach to gather individual tasks and activities, based on the
macro map, that comprise the process.
Create Level 1
4. Immediately assemble the information that has been provided
PFM into a Process Map.
5. Verify the PFM by discussing it with process owners and by
observing the actual process from beginning to end.
Define the
scope for the
Level 2 PFM
166
Process Mapping Approach
Select the
The Team Process Mapping Workshop
process 1. Add to and agree on Macro Process Map.
2. Using 8.5 X 11 paper for each macro process step, tape the
Determine
process to the wall in a linear style.
approach to 3. Process Members then list all known process tasks that they do
map the on a post-it note, one process task per note.
process
• Include the actual time spent to perform each activity, do
not include any wait time or queue time.
Complete • List any known performance data that describe the quality
Level 1 PFM
worksheet of the task.
4. Place the post-it notes on the wall under the appropriate macro
step in the order of the work flow.
5. Review process with whole group, add additional information
Create Level 1
PFM and close meeting.
6. Immediately consolidate information into a Level 2 Process
Map.
7. You will still have to verify the map by walking the process.
Define the
scope for the
Level 2 PFM
167
Steps in Generating a Level 1 PFM
Select the
Creating a Level 1 PFM
process 1. Identify a generic name for the process:
For instance: “Customer Order Process”
Determine 2. Identify the beginning and ending steps of the process:
approach to Beginning - customer calls in. Ending – baked pizza given to operations
map the
process 3. Describe the primary purpose and objective of the process (operational
definition):
The purpose of the process is to obtain telephone orders for pizzas, sell
Complete additional products if possible, let the customer know the price and
Level 1 PFM approximate delivery time, provide an accurate cook order, log the time and
worksheet
immediately give it to the pizza cooker.
4. Mentally “walk” through the major steps of the process and write them
down:
Create Level 1
PFM
Receive the order via phone call from the customer, calculate the price,
create a build order and provide the order to operations
5. Use standard flowcharting symbols to order and to illustrate the flow of
the major process steps.
Define the
scope for the
Level 2 PFM
168
Exercise – Generate a Level 1 PFM
The purpose of this exercise is to develop a Level 1 Linear Process Flow Map
Select the for the key process you have selected as your project.
process
Read the following background for the exercise: You will use your
selected key process for this exercise (if more than one person in the class is
Determine part of the same process you may do it as a small group). You may not have
approach to all the pertinent detail to correctly put together the Process Map, that is ok, do
map the
process the best you can. This will give you a starting template when you go back to
do your project. In this exercise you may use the Level 1 PFM worksheet on
Complete
the next page as an example.
Level 1 PFM
worksheet
Task 1 – Identify a generic name for the process.
Task 2 - Identify the beginning and ending steps of the process.
Task 3 - Describe the primary purpose and objective of the process
Create Level 1
PFM
(operational definition).
Task 4 - Mentally “walk” through the major steps of the process and write
them down.
Define the
Task 5 - Use standard flowcharting symbols to order and to illustrate the flow
scope for the of the major process steps.
Level 2 PFM
169
Exercise – Generate a Level 1 PFM
4. Mentally “walk” through the major steps of the process and write them down:
5. Use standard flowcharting symbols to order and to illustrate the flow of the
major process steps on a separate sheet of paper.
170
Example Template for Generating a Level 1 PFM
• Describe the primary purpose and objective of the process (operational definition):
(The purpose of the process is to obtain telephone orders for Pizzas, sell additional
products if possible, let the customer know the price and approximate delivery time,
provide an accurate cook order, log the time and immediately give it to the pizza
cooker).
• Mentally “walk” through the major steps of the process and write them down:
(Receive the order via phone call from the customer, calculate the price, create a build
order and provide the order to the chef).
• Use standard flowcharting symbols to order and to illustrate the flow of the major
process steps on a separate sheet of paper.
171
Defining the Scope of Level 2 PFM
Determine Pizza
Dough
approach to
map the No
process Take Order Add Place in Observe Check Yes Remove
1
from Cashier Ingredients Oven Frequently if Done from Oven
Start New
Complete Pizza
Level 1 PFM
worksheet Scrap
No
Tape
Pizza Place in Put on
1 Correct Box
Order on Delivery Rack
Yes Box
Create Level 1
PFM
The rules for determining the Level 2 Process Map scope:
• From your Macro Process Map, select the area which represents your problem.
• Map this area at a Level 2.
Define the
scope for the • Start and end at natural starting and stopping points for a process, in other words you
Level 2 PFM have the complete associated process.
172
Defining the Scope of Level 2 PFM
No
Take Order Add Place in Observe Check Yes Remove
Perform SIPOC from Cashier Ingredients Oven Frequently if Done from Oven 1
Start New
Pizza
Identify all X’s
and Y’s Scrap
No
Tape
Pizza Place in Put on
1 Correct Box
Order on Delivery Rack
Identify customer Yes Box
requirements
Identify supplier
requirements
173
Building a SIPOC
Customer Order:
Identify customer
requirements Level 1 process flow diagram
Call for an Answer Write Confirm Sets Address & Order to
Order Phone Order Order Price Phone Cook
Identify supplier
requirements
174
Identifying Customer Requirements
Identify customer
requirements
Identify supplier
requirements
175
Identifying Supplier Requirements
Identify customer
requirements
Identify supplier
requirements
176
Controllable vs. Noise Inputs
Screens in Place
Oven Temp
Bake Time
Controllable Key Process
Ingredients
Recipe Inputs Process Outputs
Correct Ingredients
Properly Cooked
Room Temp Pizza Size Hot Pizza >140 deg
Moisture Content
Ingredient Variation
Noise Inputs Ingredient Types/Mixes
Volume
177
Exercise – Supplier Requirements
The purpose of this exercise is to identify the requirements for the suppliers to
Create the Level
the key process you have selected as your project.
2 PFM
Read the following background for the exercise: You will use your
selected key process for this exercise (if more than one person in the class is
Perform SIPOC part of the same process you may do it as a small group). You may not have
all the pertinent detail to correctly identify all supplier requirements, that is ok,
do the best you can. This will give you a starting template when you go back
to do your workplace assignment. Use the process input identification and
Identify all X’s analysis form for this exercise.
and Y’s
178
The Level 3 Process Flow Diagram
Pizza
Dough
No
Take Order Add Place in Observe Check Yes Remove
from Cashier Ingredients Oven Frequently if Done from Oven 1
Start New
Pizza
Scrap
No
Tape
Pizza Place in Put on
1 Correct Box
Order on Delivery Rack
Yes Box
OUTPUT
1 IDENTIFICATION
3 4 AND
5 ANALYSIS
6 7 8 9 10 11 12 13 INPUT IDENTIFICATION AND ANALYSIS
1 Output Data 3 4 5 Requirements
6 Data7 8 9 Measurement
10 Data 11 Value Data
12 General Data/Information
13 INPUT
1 IDENTIFICATION
2 3 AND4 ANALYSIS
5 6 7 8 9 10 11 12 13
Customer (Name) Metric 1 Input Data2 3 4 5Requirements
6 Data7 8 9 Measurement10 Data 11 Value Data
12 General Data/Information
13
Output Data Requirements Data Measurement Measurement Data Value
VA Data General Data/Information
Customer (Name) Metric System (How is it Frequency of or VA Input Data Supplier (Name) Metric
Requirements Data Measurement Measurement Data VA Data
Value General Data/Information
Measurement Controlled (C) System (How is it Frequency of Performance or VA
Process Output - Name (Y) Internal External Metric LSL Target USL Measured) Measurement Performance Level Data NVA Comments Supplier (Name) Metric Measurement
System (How is it Frequency of or Process Input- Name (X) Noise Internal
(N) (C) External Metric LSL Target USL Measured)
Controlled System (How isMeasurement
it Frequency ofLevel Data
Performance NVA or Comments
Process Output - Name (Y) Internal External Metric LSL Target USL Measured) Measurement Performance Level Data NVA Comments
Process Input- Name (X) Noise (N) Internal External Metric LSL Target USL Measured) Measurement Level Data NVA Comments
179
Process Inputs (X’s) and Outputs (Y’s)
Process Name Step Name/Number
PROCESS STEP
OUTPUT IDENTIFICATION AND ANALYSIS
1 3 4 5 6 7 8 9 10 11 12 13
Output Data Requirements Data Measurement Data Value Data General Data/Information
Create a Level 3 Customer (Name) Metric Measurement
System (How is it Frequency of
VA
or
PFM Process Output - Name (Y) Internal External Metric LSL Target USL Measured) Measurement Performance Level Data NVA Comments
Add
Performance
data
Identify
VA/NVA steps PROCESS STEP
Process Name Step Name/Number
180
Process Inputs (X’s) and Outputs (Y’s)
A No
Greetings and Request
NV No
Call for an Answer mention order from Writes on Asks for Confirm
Order phone specials customer scratch pad more? order
Yes
1 2
A No
NV No
Calculate Asks cook Inform Gets address Thanks Another
2 price for time customer of Order
& phone # customer & call 3
estimate price/time still OK? hangs up waiting
Yes
NVA Yes
Writes time
on scratch
pad Yes
Create a Level New
1
order?
3 PFM •Each process activity can be tested for its
value-add contribution No
A
Completes NV
Add •Ask the following two questions to identify 3 order from
from note pad
182
Definition of X-Y Matrix
183
The Vital Few
A Six Sigma Belt does not just discover which X’s are important in a process
(the vital few).
– The team considers all possible X’s that can contribute or cause the
problem observed.
– The team uses 3 primary sources of X identification:
• Process Mapping
• Fishbone Analysis
• Basic Data Analysis – Graphical and Statistical
– A List of X’s is established and compiled.
– The team then prioritizes which X’s it will explore first, and eliminates
the “obvious” low impact X’s from further consideration.
184
The “X-Y Matrix”
185
Using the Classified X’s
• Breakthrough requires dealing primarily with controllable X’s impacting the “Y”.
• Use the controllable X’s from the Fishbone analysis to include in the X-Y Matrix.
• The goal is to isolate the vital few X’s from the trivial many X’s.
• Procedures and Noise X’s will be used in the FMEA at the end of this module.
However:
– All procedures must be in total compliance.
• This may require some type of effectiveness measure.
• This could reduce or eliminate some of the defects currently seen in the process
(allowing focus on controllable X’s).
– Noise type inputs increase risk of defects under current technology of
operation and therefore:
• Increase RPN on the FMEA document from an input.
• Help identify areas needing investment for a justified ROI.
186
X-Y Matrix: Steps
187
X-Y Matrix: Steps
188
X-Y Matrix: Steps
For each X listed, rank its effect on each metric based on a scale of 1, 3 or 9.
– 9 = Highest
– 3 = Marginal
– 1 = None
189
X-Y Matrix: Steps
“Ranking” multiplies the rank of each X by the Weight of each Metric. The product
of that is added together to become the “Ranking”.
190
Example
191
Example
YX Diagram Summary
Process: laminating
Date: 5/2/2006 Input Matrix Results
100.00%
Output Variables Input Variables 90.00%
80.00%
O utput (Y's)
Description Weight Description Ranking Rank % 70.00%
60.00%
broken 10 temperature 162 14.90% 50.00%
40.00%
unbonded area 9 human handling 159 14.63% 30.00%
20.00%
smears 8 material properties 130 11.96% 10.00%
0.00%
thickness 7 washer 126 11.59%
pressure
time
foreign material 6 pressure 120 11.04%
0 robot handling 120 11.04%
0 time 102 9.38%
0 clean room practices 90 8.28% Input Summary
0 clean room cleanliness 78 7.18% Input (X's)
0 - 0.00%
192
Fishbone Diagram Exercise
193
Definition of FMEA
194
History of FMEA
History of FMEA:
• First used in the 1960’s in the Aerospace industry during the Apollo
missions
• In 1974 the Navy developed MIL-STD-1629 regarding the use of FMEA
• In the late 1970’s, automotive applications driven by liability costs, began to
incorporate FMEA into the management of their processes
• Automotive Industry Action Group (AIAG) now maintains the FMEA
standard for both Design and Process FMEA’s
195
Types of FMEA’s
• System FMEA: Performed on a product or service product at the early concept/design level
when various modules all tie together. All the module level FMEA’s tie together to form a
system. As you go lower into a system more failure modes are considered.
– Example: Electrical system of a car, consists of the following modules: battery, wiring
harness, lighting control module and alternator/regulator.
– System FMEA focuses on potential failure modes associated with the modules of a system
caused by design
• Design DFMEA: Performed early in the design phase to analyze product fail modes before they
are released to production. The purpose is to analyze how fail modes affect the system and
minimize them. The severity rating of a fail mode MUST be carried into the Process PFMEA.
• Process PFMEA: Performed in the early quality planning phase of manufacturing to analyze
fail modes in manufacturing and transactional processes that may escape to the customer. The
failure modes and the potential sources of defects are rated and corrective action taken based on a
Pareto analysis ranking.
• Equipment FMEA: used to analyze failure modes in the equipment used in a process to detect
or make the part.
– Example: Test Equipment fail modes to detect open and short circuits.
196
Purpose of FMEA
FMEA’s:
• Document and track actions taken to reduce risk and improve the
process.
197
Who Creates FMEAs and When?
Who When
• The focused team working on a • Process FMEAs should be started:
breakthrough project. • At the conceptual design phase.
• Process FMEAs should be updated:
• ANYONE who had or has a role
• When an existing design or process is being
in defining, executing, or
changed.
changing the process.
• When carry-over designs or processes will be
• This includes: used in new applications and environments.
• When a problem solving study is completed
• Associates
and needs to be documented.
• Technical Experts • System FMEAs should be created after system
• Supervisors functions are defined but before specific hardware is
selected.
• Managers • Design FMEAs should be created when new
• Etc. systems, products and processes are being designed.
198
Why Create an FMEA?
As a means to manage…
RISK!!!
We want to avoid causing failures in the Process as well as the Primary &
Secondary Metrics .
199
The FMEA…
1
2
3
4
5
6
7
8
9
200
FMEA Components…#
201
FMEA Components…Process Step
Enter the Name of the Process Step here. The FMEA should sequentially follow the
steps documented in your Process Map.
Phone
Dial Number
Listen for Ring
Say Hello
Introduce Yourself
Etc.
202
FMEA Components…Potential Failure Modes
This refers to the mode in which the process could potentially fail. These are
the defects caused by a C,P or N factor that could occur in the Process.
This information is obtained from Historical Defect Data.
203
FMEA Components…Potential Failure Effects
204
FMEA Components…Severity (SEV)
This ranking should be developed based on the teams knowledge of the process in
conjunction with the predetermined scale.
The measure of Severity is a financial measure of the impact to the business of
realizing a failure in the output.
205
Ranking Severity
High Minor disruption to the production line. The product may have to be sorted and a portion 7
(less than 100%) scrapped. Vehicle operable, but at a reduced level of
performance. Customers will be dissatisfied.
Moderate Minor disruption to the production line. A portion (less than 100%) may have to be 6
scrapped (no sorting). Vehicle/item operable, but some comfort/convenience
item(s) inoperable. Customers will experience discomfort.
Low Minor disruption to the production line. 100% of product may have to be re-worked. 5
Vehicle/item operable, but some comfort/convenience item(s) operable at a
reduced level of performance. Customers will experience some dissatisfaction.
Very Low Minor disruption to the production line. The product may have to be sorted and a 4
portion (less than 100%) re-worked. Fit/finish/squeak/rattle item does not
conform. Most customers will notice the defect.
Minor Minor disruption to the production line. A portion (less than 100%) of the product may 3
have to be re-worked online but out-of-station. Fit/finish/squeak/rattle item
does not conform. Average customers will notice the defect.
Very Minor Minor disruption to the production line. A portion (less than 100%) of the product may 2
have to be re-worked online but in-station. Fit/finish/squeak/rattle item does
not conform. Discriminating customers will notice the defect.
None No effect. 1
* Potential Failure Mode and Effects Analysis (FMEA), Reference Manual, 2002. Pgs 29-45. Chrysler Corporation, Ford Motor
Company, General Motors Corporation.
206
Applying Severity Ratings to Your Process
• The guidelines presented on the previous slide were developed for the auto
industry.
• This was included only as a guideline....”actual results may vary” for your
project.
• Your severity may be linked to impact on the business or impact on the next
customer, etc.
Let’s brainstorm how we might define the following SEVERITY levels in our own
projects:
1, 5, 10
207
Sample Transactional Severities
Critical Business May endanger company’s ability to do business. Failure mode affects process
10
Unit-wide operation and / or involves noncompliance with government regulation.
Critical Loss - May endanger relationship with customer. Failure mode affects product delivered
Customer and/or customer relationship due to process failure and/or noncompliance with 9
Specific government regulation.
208
FMEA Components…Classification “Class”
Controllable – A factor that can be dialed into a specific setting/value. For example Temperature or Flow.
Procedures – A standardized set of activities leading to readiness of a step. For example Safety Compliance, “Lock -Out
Tag-Out.”
Noise - A factor that can not be dialed in to a specific setting/value. For example rain in a mine.
209
Potential Causes of Failure (X’s)
Potential Causes of the Failure refers to how the failure could occur.
This information should be obtained from the Fishbone Diagram.
210
FMEA Components…Occurrence “OCC”
211
Ranking Occurrence
Potential Failure Mode and Effects Analysis (FMEA), Reference Manual, 2002. Pg. 35.. Chrysler Corporation, Ford Motor Company,
General Motors Corporation.
212
FMEA Components…Current Process Controls
Current Process Controls refers to the three types of controls that are in place to
prevent a failure in with the X’s. The 3 types of controls are:
• SPC (Statistical Process Control)
• Poke-Yoke – (Mistake Proofing)
• Detection after Failure
213
FMEA Components…Detection (DET)
Detection is an assessment of the probability that the proposed type of control will
detect a subsequent Failure Mode.
214
Ranking Detection
Potential Failure Mode and Effects Analysis (FMEA), AIAG Reference Manual, 2002 Pg. 35.. Chrysler Corporation, Ford Motor
Company, General Motors Corporation.
215
Risk Priority Number “RPN”
The Risk Priority Number is a value that will be used to rank order the concerns
from the process.
The RPN is the product of, Severity, Occurrence and Detect ability as represented
here…
RPN = (SEV)*(OCC)*(DET)
216
FEMA Components…Actions
Responsible Person & Date refers to the name of the group or person responsible
for completing the activity and when they will complete it.
Taken Action refers to the action and effective date after it has been completed.
217
FMEA Components…Adjust RPN
Once the Recommended Actions, Responsible Person & Date, Taken Action
have been completed the Severity, Occurrence and Detection should be adjusted.
This will result in a new RPN rating.
218
FMEA Exercise
OK Team, let’s
get that FMEA!
219
Summary
• Create an FMEA
220
Six Sigma Statistics
Process Discovery
Basic Statistics
Descriptive Statistics
Normal Distribution
Assessing Normality
Graphing Techniques
Process Capability
221
Purpose of Basic Statistics
Relax….it won’t
be that bad!
222
Statistical Notation – Cheat Sheet
The Standard Deviation of population data For each, all, individual values
223
Parameters vs. Statistics
Population: All the items that have the “property of interest” under study.
Population
Sample
Sample
Sample
224
Types of Data
225
Discrete Variables
The number of times you need to flip a coin before a head 1,2,3, …
appears for the first time
(note, there is no upper limit because you might need to
flip forever before the first head appears)
226
Continuous Variables
The length of prison time served for individuals convicted All the real numbers between a and b, where a is the
of first degree murder smallest amount of time served and b is the largest.
The household income for households with incomes less All the real numbers between a and $30,000, where a is
than or equal to $30,000 the smallest household income in the population
The blood glucose reading for those individuals having All real numbers between 200 and b, where b is the
glucose readings equal to or greater than 200 largest glucose reading in all such individuals
227
Definitions of Scaled Data
• Understanding the nature of data and how to represent it can affect the types of statistical
tests possible.
• Ordinal Scale – data is arranged in some order, but differences between data values either
cannot be determined or are meaningless.
• Interval Scale – data can be arranged in some order and for which differences in data
values are meaningful. The data can be arranged in an ordering scheme and differences
can be interpreted.
• Ratio Scale – data that can be ranked and for which all arithmetic operations including
division can be performed. (division by zero is of course excluded) Ratio level data has
an absolute zero and a value of zero indicates a complete absence of the characteristic of
interest.
228
Nominal Scale
230
Interval Scale
231
Ratio Scale
232
Converting Attribute Data to Continuous Data
233
Descriptive Statistics
234
Descriptive Statistics
Enter the data as below in an Excel sheet. The data contains Average
Handle time for calls in a call center service BU
Data
5.01
5.01
5.01
5
5.01
5
5.01
4.99
4.99
5
5
5
4.99
5
5
5
5
235
Measures of Location
Mean is:
• Commonly referred to as the average.
• The arithmetic balance point of a distribution of data.
Stat>Basic Statistics>Display Descriptive Statistics…>Graphs…
>Histogram of data, with normal curve
Column1
Sample Population
Mean 5.001176
Range is the:
Difference between the largest observation and the smallest observation in the data
set.
• A small range would indicate a small amount of variability and a large range a large
amount of variability.
Sample Population
238
Measures of Variation
Variance is the:
Average squared deviation of each individual data point from the Mean.
Sample Population
239
Normal Distribution
240
The Normal Curve
241
Normal Distribution
– Has a μ = 0, and σ = 1
– Z-scores measure how many Standard Deviations from the mean a particular
data-value lies.
242
Normal Distribution
The area under the curve between any 2 points represents the proportion
of the distribution between those points.
The
Thearea
areabetween
betweenthe
theMean
Mean
and
andany
anyother
otherpoint
pointdepends
depends
upon
uponthe
theStandard
StandardDeviation.
Deviation.
x
Convert any raw score to a Z-score using the formula:
-6 -5 -4 -3 -2 -1 +1 +2 +3 +4 +5 +6
244
The Empirical Rule (cont.)
No matter what the shape of your distribution is, as you travel 3 Standard Deviations
from the Mean, the probability of occurrence beyond that point begins to converge to a
very low number.
245
Why Assess Normality?
While many processes in nature behave according to the Normal Distribution, many
processes in business, particularly in the areas of service and transactions, do not.
There are many statistical tools that assume Normal Distribution properties in their
calculations.
So understanding just how “Normal” the data are will impact how we look at the
data.
246
Tools for Assessing Normality
The shape of any Normal curve can be calculated based on the Normal
Probability density function.
Tests for Normality basically compare the shape of the calculated curve to
the actual distribution of your data points.
247
Goodness-of-Fit
100
Expected for Normal Distribution
Departure of the actual Actual Data
20%
data from the expected
80
Normal Distribution. C
u
m
The Anderson-Darling u
l
a 60
Goodness-of-Fit test t
i
v
assesses the magnitude e
P
of these departures e 40
r
using an Observed c
e
n
minus Expected t
20
formula. 20%
0
3.0 3.5 4.0 4.5 5.0 5.5
Raw Data Scale
248
The Normal Probability Plot
ENTER
DATA
HERE Sorted Count F1i 1-F1i F2i Si N Plt Line N Plt Line
Test Hypotheses
103.5 5 0.5665666 0.4334334 0.7619126 -7.5607564 103.5632 0.637931
103.5
H0: Data is normally distributed.
Is the data normally distributed or
103.7 6 0.77468896 0.22531104 non-normally distributed?
0.7619126 -5.7993879 103.7013 0.775862
103.2 HA: Data is not normally distributed.
0.679287 P-value
104 7 0.94892162 0.05107838 0.9030636 -2.0070883 103.908 0.913793
104
249
If the Data Are Not Normal, Don’t Panic!
• There are lots of meaningful statistical tools you can use to analyze your
data (more on that later).
• It just means you may have to think about your data in a slightly
different way.
251
Isolating Special Causes from Common Causes
If we know that the basic structure of the data should follow a Normal
Distribution, but plots from our data shows otherwise; we know the data contain
Special Causes.
253
Data Sources
Data sources are suggested by many of the tools that have been covered so
far:
– Process Map
– X-Y Matrix
– Fishbone Diagrams
– FMEA
Examples are:
1. Time 3. Operator
Shift Training
Day of the week Experience
Week of the month Skill
Season of the year Adherence to procedures
254
Graphical Concepts
255
The Histogram
1. IQR = 0.01, n = 17
2. Bin Width = 0.007
3. Number of Bins = 3
256
Histogram Caveat
257
Box Plot
Box Plots summarize data about the shape, dispersion and center of the data and also
help spot outliers.
Box Plots require that one of the variables, X or Y, be categorical or Discrete and the
other be Continuous.
Maximum Value
75th Percentile
Middle
50% of 50th Percentile (Median)
Data
Mean
25th Percentile
Outlier
*
Upper Limit: Q3+1.5(Q3-Q1)
Upper Whisker
Box
Q2: Median 50th Percentile
Lower Whisker
259
Box Plot Examples
Use the sheet Box Plot to draw your own Box Plots
and interpret
260
Box Plot Example
261
Time Series Plot
Time Series Plots are excellent tools and a must in all projects for you to understand if there are time-series patterns in the data.
262
Summary
• Explain the various statistics used to express location and spread of data
• Use data to generate various graphs and make interpretations based on their
output
263
Measurement System Analysis
Process Discovery
Basics of MSA
Variables MSA
Attribute MSA
Process Capability
264
Introduction to MSA
So far we have learned that the heart and soul of Six Sigma is that it is a
data-driven methodology.
– How do you know that the data you have used is accurate and precise?
– How do know if a measurement is a repeatable and reproducible?
Item to be Reference
Measured Measurement
Operator Measurement Equipment
Process
Procedure
Environment
The item to be measured can be a physical part, document or a scenario for customer service.
Operator can refer to a person or can be different instruments measuring the same products.
Reference is a standard that is used to calibrate the equipment.
Procedure is the method used to perform the test.
Equipment is the device used to measure the product.
Environment is the surroundings where the measures are performed.
266
Measurement Purpose
In order to be worth collecting, measurements must provide value - that is, they
must provide us with information and ultimately, knowledge
The question…
267
Purpose
268
Accuracy and Precision
Accurate
Accuratebut butnotnotprecise
precise--OnOnaverage,
average, Precise
Precisebut
butnot
notaccurate
accurate--TheTheaverage
average
the
theshots
shotsare
areininthe
thecenter
centerof
ofthe
thetarget
target isisnot
noton
onthe
thecenter,
center,but
butthe
thevariability
variabilityisis
but
butthere
thereisisaalot
lotofofvariability
variability small
small
269
MSA Uses
270
Why MSA?
271
Appropriate Measures
272
Poor Measures
273
Examples of What to Measure
Whenever you measure anything, the variation that you observe can be segmented
into the following components…
Observed Variation
Precision Accuracy
All measurement systems have error. If you don’t know how much of the variation you
observe is contributed by your measurement system, you cannot make confident decisions.
If you were one speeding ticket away from losing your license, how fast would you be
willing to drive in a school zone?
275
Precision
A precise metric is one that returns the same value of a given attribute
every time an estimate is made.
Precise data are independent of who estimates them or when the estimate
is made.
276
Repeatability
Repeatability
For example:
– Manufacturing: One person measures the purity of multiple samples of the same
vial and gets different purity measures.
– Transactional: One person evaluates a contract multiple times (over a period of
time) and makes different determinations of errors.
277
Reproducibility
Reproducibility
Y Operator A
Operator B
For example:
– Manufacturing: Different people perform purity test on samples from the same
vial and get different results.
– Transactional: Different people evaluate the same contract and make different
determinations.
278
Time Estimate Exercise
279
Accuracy
Accuracy
Warning, do not assume your metrology
reference is gospel.
Measurement
280
Accuracy Against a Known Standard
However, before you invest a lot of time analyzing the data, you must ensure the
data has integrity.
– The analysis should include a comparison with known reference points.
– For the example of product returns, the transaction details should add up
to the same number that appears on financial reports, such as the income
statement.
281
Accuracy vs. Precision
+ =
282
Bias
Bias is defined as the deviation of the measured value from the actual value.
Bias Bias
283
Stability
Drift
284
Linearity
Linearity is defined as the difference in Bias values throughout the measurement range in
which the gauge is intended to be used. This tells you how accurate your measurements
are through the expected range of the measurements. It answers the question, "Does my
gage have the same accuracy for all sizes of objects being measured?"
B i a s (y)
0.00
*
-e
*
*
Reference Value (x)
y = a + b.x
y: Bias, x: Ref. Value
a: Slope, b: Intercept
285
Types of MSA’s
Attribute Variable
– Pass/Fail – Continuous scale
– Go/No Go – Discrete scale
– Document Preparation – Critical dimensions
– Surface imperfections – Pull strength
– Customer Service Response – Warp
286
Variable MSA’s
MINITAB™ calculates a column of variance components (VarComp) which are used to calculate %
Gage R&R using the ANOVA Method.
Estimates for a Gage R&R study are obtained by calculating the variance components for each term
and for error. Repeatability, Operator and Operator*Part components are summed to obtain a total
Variability due to the measuring system.
We use variance components to assess the Variation contributed by each source of measurement error
relative to the total Variation.
287
Session Window Cheat Sheet
288
Session Window Cheat Sheet
289
Number of Distinct Categories
The number of distinct categories tells you how many separate groups of parts the
system is able to distinguish.
Unacceptable for estimating
process parameters and indices
Only indicates whether the
process is producing conforming
or nonconforming parts
1 Data Category
Generally unacceptable for
estimating process parameters
and indices
Only provides coarse estimates
2 - 4 Categories
Recommended
5 or more Categories
290
AIAG Standards for Gage Acceptance
Here are the Automotive Industry Action Group’s definitions for Gage
acceptance.
% Tolerance
or % Contribution System is…
% Study Variance
291
MINITABTM Graphic Output Cheat Sheet
50 0.625
0 0.620
Gage R&R Repeat Reprod Part-to-Part Part 1 2 3 4 5 6 7 8 9 10
0.630
InInaagood
goodmeasurement
measurementsystem,
0.629
system,thethelargest
largestcomponent
componentofof 3
Average
0.629
0.628 Mean=0.6282 0.628
0.627
0.626
Variation
VariationisisPart-to-Part
0.627
0.626Part-to-Partvariation.
variation. IfIfinstead
insteadyou
youhave
havelarge
large
0.625
0.624
LCL=0.6248
amounts of Variation
0.625
attributed to Gage R&R, then
amounts of Variation attributed to Gage R&R, then corrective
0.624
corrective
0
action
actionisisneeded.
Part 1 2
needed.
3 4 5 6 7 8 9 10
292
MINITABTM Graphic Output Cheat Sheet
50 0.625
0 0.620
Gage R&R Repeat Reprod Part-to-Part MINITAB1TMTMprovides
MINITAB providesanan
Part 2 3 4 5 R Chart
7 and
R6 Chart 8 Xbar
and 9 10 Chart by Operator. The R
Xbar Chart by Operator. The R
R Chart by Operator chart
chartconsists
consistsofofthe following:
By
the Operator
following:
0.010 1 2 3
0.630
- The plotted points are the difference between the largest and smallest
Sample Range
- The plotted points are the difference between the largest and smallest
UCL=0.005936
measurements on each part for each operator. If the measurements are
0.005
measurements
0.625 on each part for each operator. If the measurements are
the same then the range = 0.
R=0.001817 the same then the range = 0.
- -The
TheCenter
CenterLine,
Line,isisthe
thegrand
grandaverage
averagefor
forthe
theprocess.
process.
0.000 LCL=0 0.620
- -The Control Limits represent
The 1Control Limits represent the amount
the amount of variation expected for the
0 Operator 2 3 of variation expected for the
subgroup
subgroupranges.
ranges. These
Theselimits
limitsare
arecalculated
calculatedusing
usingthe
thevariation
variationwithin
within
Xbar Chart by Operator subgroups. Operator*Part Interaction
0.632 1 2 3
UCL=0.6316
subgroups.
0.631
Operator
0.631 1
0.630 2
Sample Mean
0.630
If any of the points on the graph go above the upper Control Limit
0.629
If any of the points on the graph go above the upper Control Limit
3
Average
0.629
0.628 Mean=0.6282 (UCL), then that operator is having problems consistently measuring
0.628
0.627 (UCL), then that operator is having problems consistently measuring
0.627
0.626 parts. The Upper Control Limit value takes into account the number of
0.626
0.625 parts. The Upper Control Limit value takes into account the number of
0.624
LCL=0.6248 measurements
measurementsby
0.625
0.624 byananoperator
operatoronona apart
partand
andthe
thevariability
variabilitybetween
betweenparts.
parts.
0 IfIf the operators are measuring consistently, then these ranges shouldbebe
the
Part operators
1 2 are
3 measuring
4 5 6 consistently,
7 8 9 10 then these ranges should
small
smallrelative
relativetotothe
thedata
dataand
andthe
thepoints
pointsshould
shouldstay
stayinincontrol.
control.
293
MINITABTM Graphic Output Cheat Sheet
50 MINITAB TM
provides an R Chart and Xbar Chart by Operator. The Xbar
MINITABTM provides an R Chart and Xbar Chart by Operator. The Xbar
0.625
Chart compares the part-to-part variation to repeatability. The Xbar chart
Chart compares the part-to-part variation to repeatability. The Xbar chart
consists of the following:
0 consists of the following:
0.620
Gage R&R Repeat Reprod Part-to-Part Part 1 2 3 4 5 6 7 8 9 10
- The plotted points are the average measurement on each part for each
R Chart by Operator - The plotted pointsBy
areOperator
the average measurement on each part for each
0.010 operator.
1 2 3
operator.
- The Center Line is the overall average for all part measurements by all
0.630
Sample Range
- The Center Line is the overall average for all part measurements by all
UCL=0.005936 operators.
0.005 operators.
- The Control Limits (UCL and LCL) are based on the variability between
0.625
- The Control Limits (UCL and LCL) are based on the variability between
R=0.001817 parts and the number of measurements in each average.
0.000 LCL=0
parts and the number of measurements in each average.
0.620
0
Because
Operator the
1 parts chosen for
2 a Gage R&R study3 should represent the
Because the parts chosen for a Gage R&R study should represent the
Xbar Chart by Operator entire range ofOperator*Part
possible parts,Interaction
this graph should ideally show lack-of-
entire range of possible parts, this graph should Operator
ideally show lack-of-
0.632 1 2 3
UCL=0.6316 control.
0.631 Lack-of-control exists when many points are1 above the Upper
0.631 control.
0.630 Lack-of-control exists when many points are above the Upper
Control Limit and/or below the Lower Control Limit.2
Sample Mean
0.630
Control Limit and/or below the Lower Control Limit.
0.629 3
Average
0.629
0.628 Mean=0.6282 0.628
0.627
0.627
In this case there are only a few points out of control which indicates the
0.626 In this case there are only a few points out of control which indicates the
0.626
0.625 LCL=0.6248 measurement system is inadequate.
0.625
0.624 measurement system is inadequate.
0.624
0 Part 1 2 3 4 5 6 7 8 9 10
294
MINITABTM Graphic Output Cheat Sheet
average measurements
average measurements taken
taken by each operator on eachpart
by each operator
Components on each
of Variation
partinin
By Part
the
thestudy,
study,arranged
arrangedby
bypart.
100
part. Each
Eachline
lineconnects
connectsthe
theaverages
averagesfor%Contribution
for 0.630
%Study Var
aasingle
singleoperator.
operator.
Percent %Tolerance
50 0.625
Ideally,
Ideally,the
thelines
lineswill
will0follow
followthe
thesame
samepattern
patternand
andthe
thepartpart 0.620
averages will vary enough
averages will vary enoughGage that
thatdifferences
R&Rdifferences
Repeat
between
between
Reprod
parts
partsare
Part-to-Part are Part 1 2 3 4 5 6 7 8 9 10
clear.
clear. R Chart by Operator By Operator
0.010 1 2 3
0.630
Sample Range
UCL=0.005936
Pattern 0.005 Means… 0.625
R=0.001817
0.000 LCL=0
Lines are virtually identical Operators are measuring the parts 0.620
0 Operator 1 2 3
the same
Xbar Chart by Operator Operator*Part Interaction
0.632 1 2 3 Operator
UCL=0.6316 0.631 1
0.631
One line is consistently higher or That operator is measuring parts 0.630 2
Sample Mean
0.630
0.629
lower than the others consistently higher or lowerMean=0.6282
than 3
Average
0.629
0.628 0.628
0.627 the others 0.627
0.626 0.626
0.625 LCL=0.6248 0.625
0.624 0.624
0 Part 1 2 3 4 5 6 7 8 9 10
Lines are not parallel or they cross The operators ability to measure a
part depends on which part is being
measured (an interaction between
operator and part)
295
MINITABTM Graphic Output Cheat Sheet
arranged
arrangedbybyoperator. Dots
Dotsrepresent
representthethemeasurements;
measurements;the
%Tolerance
Percent
operator.
50 the
circle-cross
circle-crosssymbols
symbolsrepresent
representthe
themeans.
means. The
Thered
redline
line
0.625
UCL=0.005936
0.005
0.625
Parallel to the x-axis The operators are R=0.001817
0.000 measuring the parts LCL=0 0.620
0 similarly Operator 1 2 3
0.630
part 0.629 3
Average
0.629
measurement
measurement is the same using this graph. Is the spreadininthe
Not parallelis
tothe
thesame using
x-axis0.628 this
The graph. Is
operators the spread
are Mean=0.6282
the
0.628
0.627 0.627
measurements similar? Or is one operator
measuring more
the Variable
parts
measurements similar? Or is one operator more Variable
0.626 than
than 0.626
0.625 LCL=0.6248 0.625
the others? 0.624 differently
the others? 0
0.624
Part 1 2 3 4 5 6 7 8 9 10
296
MINITABTM Graphic Output Cheat Sheet
50
average measurements for each part. 0.625
0 0.620
Ideally, Gage R&R Repeat Reprod Part-to-Part Part 1 2 3 4 5 6 7 8 9 10
Multiple measurements
R Chart byfor each individual
Operator By Operator
part have little variation (the dots for one
0.010 1 2 3
part 0.630
Sample Range
0.630
0.629 3
Average
0.629
0.628 Mean=0.6282 0.628
0.627 0.627
0.626 0.626
0.625 LCL=0.6248 0.625
0.624 0.624
0 Part 1 2 3 4 5 6 7 8 9 10
297
Practical Conclusions
For this example, the measuring system contributes a great deal to the overall Variation, as confirmed
by both the Gage R&R table and graphs.
The Variation due to the measurement system, as a percent of study Variation is causing 92.21% of
the Variation seen in the process.
By AIAG Standards this gage should not be used. By all standards, the
data being produced by this gage is not valid for analysis.
% Tolerance
or % Contribution System is…
% Study Variance
298
Repeatability and Reproducibility Problems
Repeatability Problems:
• Calibrate or replace gage.
• If only occurring with one operator, re-train.
Reproducibility Problems:
• Measurement machines
– Similar machines
• Ensure all have been calibrated and that the standard measurement method is being
utilized.
– Dissimilar machines
• One machine is superior.
• Operators
– Training and skill level of the operators must be assessed.
– Operators should be observed to ensure that standard procedures are followed.
• Operator/machine by part interactions
– Understand why the operator/machine had problems measuring some parts and not others.
• Re-measure the problem parts
• Problem could be a result of gage linearity
• Problem could be fixture problem
• Problem could be poor gage design
299
Design Types
Crossed Design
• A Crossed Design is used only in non-destructive testing and assumes that all the parts can be
measured multiple times by either operators or multiple machines.
– Gives the ability to separate part-to-part Variation from measurement system Variation.
– Assesses Repeatability and Reproducibility.
– Assesses the interaction between the operator and the part.
Nested Design
• A Nested Design is used for destructive testing (we will learn about this in MBB training) and also
situations where it is not possible to have all operators or machines measure all the parts multiple
times.
– Destructive testing assumes that all the parts within a single batch are identical enough to claim
they are the same.
– Nested designs are used to test measurement systems where it is not possible (or desirable) to
send operators with parts to different locations.
– Do not include all possible combinations of factors.
– Uses slightly different mathematical model than the Crossed Design.
300
Gage R & R Study
301
Variable Gage R & R Steps
Step 1: Call a team meeting and introduce the concepts of the Gage R&R
Step 2: Select parts for the study across the range of interest
– If the intent is to evaluate the measurement system throughout the process range, select parts
throughout the range
– If only a small improvement is being made to the process, the range of interest is now the
improvement range
Step 3: Identify the inspectors or equipment you plan to use for the analysis
– In the case of inspectors, explain the purpose of the analysis and that the inspection system is
being evaluated not the people
Step 4: Calibrate the gage or gages for the study
– Remember Linearity, Stability and Bias
Step 5: Have the first inspector measure all the samples once in random order
Step 6: Have the second inspector measure all the samples in random order
– Continue this process until all the operators have measured all the parts one time
– This completes the first replicate
Step 7: Repeat steps 5 and 6 for the required number of replicates
– Ensure there is always a delay between the first and second inspection
Step 8: Enter the data into MINITABTM and analyze your results
Step 9: Draw conclusions and make changes if necessary
302
Gage R & R Study
Trial 1
Operator 1
Trial 2
P
a
Trial 1
r 1 2 3 4 5 6 7 8 9 10 Operator 2
t Trial 2
s
Trial 1
Operator 3
Trial 2
303
Data Collection Sheet
304
The Data Collection Sheet
305
Gage R & R
Variables:
– Part
– Operator
– Response
306
Gage R & R
307
Graphical Output
Looking at the “Components of Variation” chart, the Part to Part Variation needs to be larger than Gage
Variation.
If in the “Components of Variation” chart the “Gage R&R” bars are larger than the “Part-to-Part” bars, then
all your measurement Variation is in the measuring tool i.e.… “maybe the gage needs to be replaced”. The
same concept applies to the “Response by Operator” chart. If there is extreme Variation within operators,
then the training of the operators is suspect.
needs to be larger 50
0.75
0.75
0.05 _
R=0.0333
0.50
0.00 LCL=0
1 2 3
Operator
Xbar Chart by Operator
1 2 3 Operator * Part Interaction
1.00 1.00 O perator
Sample Mean
1
_
UCL=0.8774
Average
2
X=0.8147
0.75 LCL=0.7520 0.75 3
0.50
0.50
1 2 3 4 5 6 7 8 9 10
Part
308
Session Window
310
Signal Averaging
Signal Averaging can be used to reduce Repeatability error when a better gage is
not available.
– Uses average of repeat measurements.
– Uses Central Limit theorem to estimate how many repeat measures are
necessary.
311
Signal Averaging Example
Suppose the Standard Deviation for one part measured by one person many times is
9.5.
312
Signal Averaging Example
Using
Usingthe
theaverage
averageofof66repeated
repeated
measures
measureswill
willreduce
reducethethe
Repeatability
Repeatabilitycomponent
componentof of
measurement
measurementerror
errortotothe
the
desired
desired15%
15%level.
level.
1. Cut a piece of paper into 12 different lengths that are all fairly close to
one another but not too uniform. Label the back of the piece of paper to
designate its “part number”
2. Perform a variable gage R&R study as outlined in this module. Use the
following guidelines:
– Number of parts: 12
– Number of inspectors: 3
– Number of trials: 5
3. Create a MINITABTM data sheet and enter the data into the sheet as each
inspector performs a measurement. If possible, assign one person to
data collection.
4. Analyze the results and discuss with your mentor.
314
Attribute MSA
Attribute
Attribute Gage
Gage Error
Error
Repeatability
Repeatability Reproducibility
Reproducibility Calibration
Calibration
– They are used in situations where a continuous measure cannot be obtained.
– It requires a minimum of 5x as many samples as a continuous study.
– Disagreements should be used to clarify operational definitions for the
categories.
• Attribute data are usually the result of human judgment (which category does this
item belong in).
• When categorizing items (good/bad; type of call; reason for leaving) you need a
high degree of agreement on which way an item should be categorized.
315
Attribute MSA Purpose
An Attribute MSA is similar in many ways to the continuous MSA, including the purposes. Do you
have any visual inspections in your processes? In your experience how effective have they been?
316
Visual Inspection Test
Take 60 Seconds and count the number of times “F” appears in this paragraph?
The Necessity of Training Farm Hands for First Class Farms in the
Fatherly Handling of Farm Live Stock is Foremost in the Eyes of
Farm Owners. Since the Forefathers of the Farm Owners Trained the
Farm Hands for First Class Farms in the Fatherly Handling of Farm
Live Stock, the Farm Owners Feel they should carry on with the
Family Tradition of Training Farm Hands of First Class Farmers in
the Fatherly Handling of Farm Live Stock Because they Believe it is
the Basis of Good Fundamental Farm Management.
317
How can we Improve Visual Inspection?
Date of study :
Assessment Agreement
Reported by :
Name of product:
Misc:
A ppraiser vs Standard
100 95.0% C I
P ercent
80
60
Percent
40
20
0
Duncan Hayes Holmes Montgomery Simpson
Appraiser
This graph shows how each appraiser compared to the right answer, accuracy. The blue
dot is the actual percentage for each operator. The red line with the X on each end is the
confidence interval. Duncan agreed with the standard 53% of the time. We are 95%
confident, based on this study, that Duncan will agree with the standard between 27% and
79% of the time. To decrease the interval, add more parts to the study.
320
Attribute Agreement Analysis
Assessment Agreement
# Matched: Appraiser's assessment across trials agrees with the known standard.
321
Attribute Agreement Analysis
Between Appraisers
Assessment Agreement
Assessment Agreement
323
M&M Exercise
2 M&M Fail
• Pick 50 M&Ms out of a package.
3 M&M Pass
• Enter results into either the Excel template or MINITAB TM and draw
conclusions.
• The instructor will represent the customer for the Attribute score.
324
Summary
325
Process Capability
Process Discovery
Process Capability
Continuous Capability
Concept of Stability
Attribute Capability
326
Understanding Process Capability
Process Capability:
• The inherent ability of a process to meet the expectations of the customer without any
additional efforts.
327
Capability as a Statistical Problem
Define a Practical
Problem
Create a
Statistical Problem
Correct the
Statistical Problem
328
Capability Analysis
Frequency
Op i Verified Op i + 1
? Data for
Y1…Yn
X1
Y1 10.16
9.87
X2 Off-Line
10.11
10.16 10.16
Analysis Scrap 10.05
10.11 9.99
9.87 10.11
10.12
9.99 10.05
Correction 10.33
10.05
10.44
10.33 10.43
10.12 10.33
X3 Y2 9.86
10.44 10.21
10.43 10.44
10.01
10.21 9.86
9.80 9.90 10.0 10.1 10.2 10.3 10.4 10.5
10.07
9.86
10.29
10.07 10.15
10.01 10.07
10.36
10.29 10.44
10.15 10.29
X4 10.36 10.03
10.44 10.36
10.33
10.03
10.15
10.33
Yes No Y3 10.15
X5 Correctable
?
Requirements – “Voice
Critical X(s): Data - VOP of the Customer”
Any variable(s) 10.16
10.11 9.87 10.16
LSL = 9.96 USL = 10.44
10.05 9.99 10.11
undue influence on
9.86 10.44
10.07 10.01 9.86
10.29 10.15 10.07
Defects
process Defects
9.70 9.80 9.90 10.0 10.1 10.2 10.3 10.4 10.5 10.6
Compares the VOP to the VOC Percent Composition
329
Process Output Categories
Target Target
Re
ss
du
Capable and e
oc
ce
on target r
rp
sp
te
re
Average n
ad
LSL USL Ce
Target
330
Problem Solving Options – Shift the Mean
This involves finding the variables that will shift the process over to the target.
This is usually the easiest option.
USL
LSL
Shift
331
Problem Solving Options – Reduce Variation
This is typically not so easy to accomplish and occurs often in Six Sigma
projects.
LSL USL
332
Problem Solving Options – Shift Mean & Reduce Variation
USL
LSL Shift & Reduce
333
Problem Solving Options
Obviously this implies making them wider, not narrower. Customers usually
do not go for this option but if they do…it’s the easiest!
334
Capability Studies
Capability Studies:
• Are intended to be regular, periodic, estimations of a process’s ability to meet its
requirements.
• Can be conducted on both Discrete and Continuous Data.
• Are most meaningful when conducted on stable, predictable processes.
• Are commonly reported as Sigma Level which is optimal (short term)
performance.
• Require a thorough understanding of the following:
– Customer’s or business’s specification limits
– Nature of long-term vs. short-term data
– Mean and Standard Deviation of the process
– Assessment of the Normality of the data (Continuous Data only)
– Procedure for determining Sigma level
335
Steps to Capability
#1 Verify Customer
Requirements
#2 Validate
Specification
Limits
#3 Collect Sample
Data
#4 Determine
Data Type
(LT or ST)
#5 Check data
for normality
#6 Calculate
Z-Score, PPM,
Yield, Capability
Cp, Cpk, Pp, Ppk
#7
336
Verifying the Specifications
Questions to consider:
337
Data Collection
Capability Studies should include “all” observations (100% sampling) for a specified period.
Short-term data: Long-term data:
•Collected across a narrow inference space. •Is collected across a broader inference space.
•Daily, weekly; for one shift, machine, •Monthly, quarterly; across multiple shifts,
operator, etc. machines, operators, etc
•Is potentially free of special cause •Subject to both common and special causes of
variation. variation.
•Often reflects the optimal performance •More representative of process performance over
level. a period of time.
•Typically consists of 30 – 50 data points. •Typically consists of at least 100 – 200 data
points.
F ill Q u a n tity
Lot 1 Lot 5
Lot 3
Lot 2
Lot 4
Short-term studies
Long-term study
338
Baseline Performance
Process
ProcessBaseline:
Baseline:The
Theaverage,
average,
long-term
long-termperformance
performancelevel
levelof
ofaa
process
processwhen
whenall
allinput
inputvariables
variablesare
are
unconstrained.
unconstrained. Long-term
Long-term
baseline
baseline
Short
4
ShortTerm
Term
Performance
Performance
` 3
2
1
LSL TARGET USL
339
Components of Variation
Even stable processes will drift and shift over time by as much as 1.5 Standard
Deviations on the average.
Long Term
Overall Variation
Short Term
Between Group Variation
Short Term
Within Group Variation
340
Sum of the Squares Formulas
Precision
Shift (short-term capability)
x
Output Y
x x
x
x x
x x x
x x
x
x x x Time
x x
x x x x
x x x
x
341
Stability
A Stable Process is consistent over time. Time Series Plots and Control Charts are
the typical graphs used to determine stability.
At this point in the Measure Phase there is no reason to assume the process is stable.
Tic toc…
tic toc…
342
Measures of Capability
Mathematically Cpk and Ppk are the same and Cp and Pp are the same.
The only difference is the source of the data, Short-term and Long-term, respectively.
– Cp and Pp
• What is Possible if your process is perfectly Centered Hope
• The Best your process can be
• Process Potential (Entitlement)
343
Capability Formulas
Sample Mean
Note: Consider the “K” value the penalty for being off center LSL – Lower specification limit
USL – Upper specification limit
344
MINITAB™ Example
345
MINITAB™ Example
Assuming the same sample data set in the previous exercise, let us find
out the Capability indices of the data set.
Standard
LSL USL Xbar Deviation Cp Cpl Cpu Cpk
4.94 5.84 5.12 0.17 0.87 0.35 1.40 0.35
346
Process Performance
Zst 2.61
Zlt 1.04
Zst or Short Term Sigma is calculated with the
formula, Cp * 3.
1. Sub group sizes will alter the way how Cp and Cpk are calculated.
2. With subgroup sizes coming into play, variations below come into play
a. Within Sub group variations
b. Between Sub group variations
347
Capability Steps
#3 Collect Sample
Data
#4 Determine
Data Type
(LT or ST)
#5 Check data
for Normality
Calculate
#6 Z-Score, PPM,
Yield, Capability
Cp, Cpk, Pp, Ppk
#7
348
Attribute Capability Steps
Validate
#2 Specification
Limits
#3 Collect Sample
Data
#4
Calculate
DPU
#5
Find Z-Score
#6 Convert Z-Score
to Cp & Cpk
#7
349
Z Scores
The Z Score effectively transforms the actual data into standard normal units. By
referring to a standard Z table you can estimate the area under the Normal curve.
– Given an average of 50 with a Standard Deviation of 3 what is the proportion
beyond the upper spec limit of 54?
50
54
350
Z Table
351
Attribute Capability
Attribute data is always long-term in the shifted condition since it requires so many samples
to get a good estimate with reasonable confidence.
1 158655.3 691462.5
Short Term Subtract
ZST Capability 1.5 2 22750.1 308537.5
3 1350.0 66807.2
Long Term Add
ZLT Capability 1.5
4 31.7 6209.7
5 0.3 232.7
By viewing these formulas you can see there is a relationship between them.
If we divide our Z short-term by 3 we can determine our Cpk and if we divide our Z long-
term by 3 we can determine our Ppk.
353
Attribute Capability Example
A total of 20,000 calls came in during the month but 2,500 of them “dropped” before
they were answered (the caller hung up).
1. Calculate DPU
2. Look up DPU value on the Z-Table
3. Find Z-Score
4. Convert Z Score to Cpk, Ppk
Example:
Look up ZLT
ZLT = 1.11
Convert ZLT to ZST = 1.11+1.5 = 1.61
355
Attribute Capability
1. Calculate DPU
2. Look up DPU value on the Z-Table
3. Find Z Score
4. Convert Z Score to Cpk, Ppk
Example:
Look up ZLT
ZLT = 1.11
Convert ZLT to ZST = 1.11+1.5 = 1.61
356
Summary
357
Measure Phase
Wrap Up and Action Items
Measure Phase Overview - The Goal
359
Six Sigma Behaviors
Listed below are the Measure Deliverables that each candidate should present in a
Power Point presentation to their mentor and project champion.
At this point you should understand what is necessary to provide these deliverables
in your presentation.
– Team Members (Team Meeting Attendance)
– Primary Metric
– Secondary Metric(s)
– Process Map – detailed
– FMEA
– X-Y Matrix
– Basic Statistics on Y
– MSA
– Stability graphs
– Capability Analysis
– Project Plan
– Issues and Barriers
361
Measure Phase - The Roadblocks
Look for the potential roadblocks and plan to address them before they become
problems:
– Team members do not have the time to collect data.
– Data presented is the best guess by functional managers.
– Process participants do not participate in the creation of the X-Y Matrix,
FMEA and Process Map.
It won’t all be
smooth sailing…..
362
DMAIC Roadmap
Process Owner
Champion/
Estimate COPQ
Establish Team
Measure
363
Measure Phase
Select the Vital Few X’s Causing Problems (X-Y Matrix, FMEA)
Y
Repeatable &
Reproducible?
N
364
Measure Phase Checklist
Measure Questions
Identify critical X’s and potential failure modes
• Is the “as is” Process Map created?
• Are the decision points identified?
• Where are the data collection points?
• Is there an analysis of the measurement system?
• Where did you get the data?
Identify critical X’s and potential failure modes
• Is there a completed X-Y Matrix?
• Who participated in these activities?
• Is there a completed FMEA?
• Has the Problem Statement changed?
• Have you identified more COPQ?
Stability Assessment
• is the “Voice of the Process” stable?
• If not, have the special causes been acknowledged?
• Can the good signals be incorporated into the process?
• Can the bad signals be removed from the process?
• How stable can you make the process?
Capability Assessment
• What is the short-term and long-term Capability of the process?
• What is the problem, one of centering, spread or some combination?
General Questions
• Are there any issues or barriers that prevent you from completing this phase?
• Do you have adequate resources to complete the project?
365
Planning for Action
366
Summary
367
Analyze Phase
Inferential Statistics
Welcome to Analyze
Inferential Statistics
Hypothesis Testing ND P1
Hypothesis Testing ND P2
369
Nature of Inference
So many
questions….?
371
Types of Error
1. Error in sampling
– Error due to differences among samples drawn at random from the population
(luck of the draw).
– This is the only source of error that statistics can accommodate.
2. Bias in sampling
– Error due to lack of independence among random samples or due to systematic
sampling procedures (height of horse jockeys only).
3. Error in measurement
– Error in the measurement of the samples (MSA/GR&R).
4. Lack of measurement validity
– Error in the measurement does not actually measure what it intends to measure
(placing a probe in the wrong slot measuring temperature with a thermometer that
is just next to a furnace).
372
Population, Sample, Observation
Population
– EVERY data point that has ever been or ever will be generated from a given
characteristic.
Sample
– A portion (or subset) of the population, either at one time or over time.
X
X X
X X
Observation
– An individual measurement.
373
Significance
Significance is all about differences. In general, larger differences (or deltas) are
considered to be “more significant.”
Practical difference and significance is:
– The amount of difference, change, or improvement that will be of practical,
economic or technical value to you.
– The amount of improvement required to pay for the cost of making the
improvement.
Statistical difference and significance is:
– The magnitude of difference or change required to distinguish between a true
difference, change or improvement and one that could have occurred by chance.
Six Sigma decisions will ultimately have a return on resource investment
(RORI)* element associated with them.
– The key question of interest for our decisions “is the benefit of making a change
worth the cost and risk of making it?”
* RORI includes not only dollars and assets but the time and participation of your teams.
374
The Mission
Variation
Mean Shift Both
Reduction
Your mission, which you have chosen to accept, is to reduce cycle time, reduce the error rate, reduce costs, reduce
investment, improve service level, improve throughput, reduce lead time, increase productivity… change the output
metric of some process, etc…
In statistical terms, this translates to the need to move the process Mean and/or reduce the process Standard Deviation.
You’ll be making decisions about how to adjust key process input variables based on sample data, not population data
- that means you are taking some risks.
How will you know your key process output variable really changed, and is not just an unlikely sample? The Central
Limit Theorem helps us understand the risk we are taking and is the basis for using sampling to estimate population
parameters.
375
A Distribution of Sample Means
Imagine you have some population. The individual values of this population form
some distribution.
Take a sample of some of the individual values and calculate the sample Mean.
The Central Limit Theorem says that as the sample size becomes large, this new
distribution (the sample Mean distribution) will form a Normal Distribution, no
matter what the shape of the population distribution of individuals.
376
Different Distributions
We will draw samples of 3 each and draw 10 samples. Let us see how
Central Limit Theorem works:
Data Sample Mean
5.01
5.01
5.01 5.01
5
5.01
5 5.00
5.01
4.99
4.99 5.00
5
5
5 5.00
4.99
5
1. Find out the average of each sample
5
5
5.00
2. Find out the grand average of all the
5
5.15 5.05 averages.
5.16
5.19
5.19
5.18
3. This is known as Mean of Means, and is
5.2
5.32
almost close to Population Mean
5.34 5.29
5.35
5.4
5.42 5.39
5.6
5.42
5.36 5.46
Put together, the three primary clauses will help us deal with inferential
statistics with ease.
378
Observations
As the sample size (number of dice) increases from 1 to 5 to 10, there are three
points to note:
1. The Center remains the same.
2. The variation decreases.
3. The shape of the distribution changes - it tends to become Normal.
The Mean of the sample Mean The Standard Deviation of the sample
distribution: Mean distribution, also known as the
Standard Error.
Good news: the Mean of the sample Mean Better news: I can reduce my uncertainty
distribution is the Mean of the population. about the population Mean by increasing my
sample size n.
379
Central Limit Theorem
If all possible random samples, each of size n, are taken from any population
with a Mean μ and Standard Deviation σ, the distribution of sample Means will:
have a Mean
Bigger is Better!
380
So What?
381
A Practical Example
Let’s say your project is to reduce the setup time for a large casting:
382
Sample Size and the Mean
383
Standard Error of the Mean
The Standard Deviation for the distribution of Means is called the standard
error of the Mean and is defined as:
– This formula shows that the Mean is more stable than a single observation by a
factor of the square root of the sample size.
384
Standard Error
The rate of change in the Standard Error approaches zero at about 30 samples.
Standard Error
0 5 10 20 30
Sample Size
This is why 30 samples is often recommended when generating summary
statistics such as the Mean and Standard Deviation.
This is also the point at which the t and Z distributions become nearly equivalent.
385
Summary
386
Hypothesis Testing (ND)
Hypothesis Testing ND P2
387
Six Sigma Goals and Hypothesis Testing
Our goal is to improve our Process Capability, this translates to the need to move the process Mean (or
proportion) and reduce the Standard Deviation.
– Because it is too expensive or too impractical (not to mention theoretically impossible) to collect
population data, we will make decisions based on sample data.
– Because we are dealing with sample data, there is some uncertainty about the true population
parameters.
Hypothesis Testing helps us make fact-based decisions about whether there are different population
parameters or that the differences are just due to expected sample variation.
96 100 104 108 112 116 120 102 105 108 111 114 117 120
O bserv ed P erformance E xp. Within P erformance E xp. O v erall P erformance O bserv ed P erformance E xp. Within P erformance E xp. O v erall P erformance
P P M < LS L 6666.67 P P M < LS L 115.74 P P M < LS L 55078.48 P P M < LS L 0.00 P P M < LS L 0.00 P P M < LS L 0.00
P P M > U S L 0.00 P P M > U S L 0.71 P P M > U S L 18193.49 P P M > U S L 0.00 P P M > U S L 0.00 P P M > U S L 0.00
P P M Total 6666.67 P P M T otal 116.45 P P M T otal 73271.97 P P M Total 0.00 P P M Total 0.00 P P M Total 0.00
388
Purpose of Hypothesis Testing
The purpose of appropriate Hypothesis Testing is to integrate the Voice of the Process with
the Voice of the Business to make data-based decisions to resolve problems.
Hypothesis Testing can help avoid high costs of experimental efforts by using existing data.
This can be likened to:
– Local store costs versus mini bar expenses.
– There may be a need to eventually use experimentation, but careful data analysis
can indicate a direction for experimentation if necessary.
389
The Basic Concept for Hypothesis Tests
Recall from the discussion on classes and cause of distributions that a data set
may seem Normal, yet still be made up of multiple distributions.
0.8
0.8
0.7
0.7
0.6
0.6
0.5
0.5
freq
0.4
0.4
0.3
0.3
0.2
0.2
0.1
0.1
0.0
0.0
-3
-3 -2
-2 -1
-1 0
0 1
1 22 3
3
xx
Are the two distributions “significantly” different from each other? How
sure are we of our decision?
Sample 1 Sample 2
391
Detecting Significance
392
Practical vs. Statistical
393
Detecting Significance
394
Hypothesis Testing
395
DICE Example
We could throw it a number of times and track how many each face occurred. With a
standard die, we would expect each face to occur 1/6 or 16.67% of the time.
If we threw the die 5 times and got 5 one’s, what would you conclude? How sure can you
be?
– Pr (1 one) = 0.1667 Pr (5 ones) = (0.1667)5 = 0.00013
There are approximately 1.3 chances out of 1000 that we could have gotten 5 ones with a
standard die.
Therefore, we would say we are willing to take a 0.1% chance of being wrong about our
hypothesis that the die was “loaded” since the results do not come close to our predicted
outcome.
396
Hypothesis Testing
DECISIONS
β n
397
Statistical Hypotheses
398
Steps to Statistical Hypothesis Test
Any differences between observed data and claims made under H 0 may be real
or due to chance.
The a level of a test (level of significance) represents the yardstick against which
P-values are measured and H0 is rejected if the
P-value is less than the alpha level.
The most commonly used levels are 5%, 10% and 1%.
400
Hypothesis Testing Risk
The alpha risk or Type 1 Error (generally called the “Producer’s Risk”) is the
probability that we could be wrong in saying that something is “different.” It is an
assessment of the likelihood that the observed difference could have occurred by
random chance. Alpha is the primary decision-making tool of most statistical tests.
Actual Conditions
Not Different Different
(Ho is True) (Ho is False)
401
Alpha Risk
Region of Region of
DOUBT DOUBT
402
Hypothesis Testing Risk
The beta risk or Type 2 Error (also called the “Consumer’s Risk”) is the
probability that we could be wrong in saying that two or more things are the same
when, in fact, they are different.
Actual Conditions
Not Different Different
(Ho is True) (Ho is False)
403
Beta Risk
Beta Risk is the probability of failing to reject the null hypothesis when a
difference exists.
Distribution if H0 is true
Reject H0
= Pr(Type 1 error)
= 0.05
H0 value
Critical
Critical value
value ofof test
test
statistic
statistic
404
Distinguishing between Two Samples
Theoretical Distribution
When n=30, we can distinguish of Means
between the Means (< 5% overlap, P- When n = 30
= 5
value < 0.05) There is a significant S=1
difference.
405
Delta Sigma—The Ratio between and S
406
Typical Questions on Sampling
407
The Perfect Sample Size
40 50 60 70
408
Hypothesis Testing Roadmap
Normal
uous
n
onti ta
C Da
409
Hypothesis Testing Roadmap
uous
n
onti ta
C Da Non Normal
410
Hypothesis Testing Roadmap
Minitab:
Stat - Tables - Chi-Square Test
If P-value < 0.05 the factors are not
independent
411
Common Pitfalls to Avoid
While using Hypothesis Testing the following facts should be borne in mind at the
conclusion stage:
– The decision is about Ho and NOT Ha.
– The conclusion statement is whether the contention of Ha was upheld.
– The null hypothesis (Ho) is on trial.
– When a decision has been made:
• Nothing has been proved.
• It is just a decision.
• All decisions can lead to errors (Types I and II).
– If the decision is to “Reject Ho,” then the conclusion should read “There is sufficient
evidence at the α level of significance to show that “state the alternative hypothesis
Ha.”
– If the decision is to “Fail to Reject Ho,” then the conclusion should read “There isn’t
sufficient evidence at the α level of significance to show that “state the alternative
hypothesis.”
412
Summary
413
Hypothesis Testing Normal Data Part 1
Inferential Statistics
Analyzing Results
Hypothesis Testing ND P2
414
Test of Means (t-tests)
Target μsample
MINITAB TM performs a one sample t-test or t-confidence interval for the Mean.
Use 1-sample t to compute a confidence interval and perform a Hypothesis Test of the
Mean when the population Standard Deviation, σ, is unknown. For a one or two-tailed 1-
sample t:
416
1 Sample t-test Sample Size
Target
Population
S
SE Mean
n
417
Sample Size
Use the 1- Sample t test Excel template provided to you in the Data Set. Enter your sample data in the excel sheet and the results to be seen as below
1.You wish to understand if your sample data is good enough to meet a target
2.You wish to understand if in the past your sample data has indeed met the target
Hypothesis
Null Hypothesis: The Hypothesized Mean/ Population Mean is the same as the sample mean statistically
Alternate Hypothesis: Hypothesized Mean/ Population Mean is not the same as the sample mean statistically
Sample Size 6
α = 0.05.
Population Mean 5.5
If p value is less than 0.05, reject
Sample Mean 12.5
the null hypothesis, else fail to
Standard Deviation 1.870828693
reject the null hypothesis.
Standard Error 0.763762616
t 9.16515139
df 5
418
Sample Size
1-Sample t Test
1. Practical Problem:
• We are considering changing suppliers for a part that we currently purchase from a
supplier that charges us a premium for the hardening process.
• The proposed new supplier has provided us with a sample of their product. They have
stated that they can maintain a given characteristic of 5 on their product.
• We want to test the samples and determine if their claim is accurate.
2. Statistical Problem:
Ho: μN.S. = 5
Ha: μN.S. ≠ 5
420
Example
4. Sample Size:
• Open the MINITABTM worksheet: “Exh_Stat.MTW”.
• Use the C1 column: Values
– In this case, the new supplier sent 9 samples for
evaluation.
– How much of a difference can be detected with this
sample?
421
1-Sample t Example
422
Example: Follow the Road Map
60
50
40
30
20
Are the data in
10
the values
column Normal?
423
1-Sample t Example
Click “Graphs”
-Select all 3
Click “Options…”
- In CI enter ’95’
424
Histogram of Values
Histogram
Histogramof
of Values
Values
(with
(with Ho
Ho and
and 95%
95%t-confidence
t-confidence interval
interval for
for the
the mean)
mean)
2.0
2.0
1.5
1.5
Frequency
Frequency
1.0
1.0
0.5
0.5
0.0
0.0 __
XX
Ho
Ho
4.4
4.4 4.5
4.5 4.6
4.6 4.7
4.7 4.8
4.8 4.9
4.9 5.0
5.0 5.1
5.1
Values
Values
Note our target Mean (represented by red Ho) is outside our population
confidence boundaries which tells that there is significant difference
between population and target Mean.
425
Box Plot of Values
Boxplot
Boxplot of
of Values
Values
(with
(with Ho
Ho and
and 95%
95%t-confidence
t-confidence interval
interval for
for the
the mean)
mean)
__
XX
Ho
Ho
4.4
4.4 4.5
4.5 4.6
4.6 4.7
4.7 4.8
4.8 4.9
4.9 5.0
5.0 5.1
5.1
Values
Values
426
Individual Value Plot (Dot Plot)
Individual
Individual Value
Value Plot
Plot of
of Values
Values
(with
(with Ho
Ho and
and 95%
95% t-confidence
t-confidence interval
interval for
for the
the mean)
mean)
__
XX
Ho
Ho
4.4
4.4 4.5
4.5 4.6
4.6 4.7
4.7 4.8
4.8 4.9
4.9 5.0
5.0 5.1
5.1
Values
Values
427
Session Window
Ho Ha
n
(X i X) 2
One-Sample T: Values
s
i 1 n 1
Test of mu = 5 vs not = 5 S
SE Mean
n
N – sample size
Mean – calculate mathematic average
StDev – calculated individual Standard Deviation (classical method)
SE Mean – calculated Standard Deviation of the distribution of the Means
Confidence Interval that our population average will fall between 4.5989, 4.9789
428
Evaluating the Results
Since the P-value of 0.034 is less than 0.05, reject the null hypothesis.
Based on the samples given there is a difference between the average of the sample
and the desired target.
X Ho
429
Manual Calculation of 1- Sample t
430
Manual Calculation of 1- Sample t
degrees of T - Distribution
freedom
.600 .700 .800 .900 .950 .975 .990 .995
1 0.325 0.727 1.376 3.078 6.314 12.706 31.821 63.657
2 0.289 0.617 1.061 1.886 2.920 4.303 6.965 9.925
3 0.277 0.584 0.978 1.638 2.353 3.182 4.541 5.841
4 0.271 0.569 0.941 1.533 2.132 2.776 3.747 4.604
5 0.267 0.559 0.920 1.476 2.015 2.571 3.365 4.032
-2.306 2.306
0
If the calculated t-value lies anywhere Critical Regions
in the critical regions, reject the null hypothesis.
– The data supports the alternative hypothesis that the estimate for the
Mean of the population is not 5.0.
431
Confidence Intervals for Two-Sided t-test
s s
X t α/2, n 1 μ X t α/2, n 1
n n
or
X t crit SE mean 4.788 2.306 * .0824
4.5989 to 4.9789
4.5989 X Ho
4.9789
4.7889
432
1-Sample t Exercise
3. Are we on Target?
433
1-Sample t Exercise: Solution
Since we do not know the population Standard Deviation, we will use the 1 sample T test to
determine if we are at Target.
434
1-Sample t Exercise: Solution
435
1-Sample t Exercise: Solution
Frequency
4
436
1-Sample t Exercise: Solution
In MINITABTM’s Session Window (ctrl – M), you can see the P-value of 0.201.
Because it is above 0.05, we “fail to reject” the null hypothesis so we accept the
equipment is giving product at a target of 32.0 ppm VOC.
437
Hypothesis Testing Roadmap
Normal
uous
n
o nti ta
C Da
438
2 Sample t-test
Use 2-Sample t to perform a Hypothesis Test and compute a confidence interval of the
difference between two population Means when the population Standard Deviations, σ’s,
are unknown.
1 2
439
Sample Size
The first step to any 2 Sample tests for means is to know if the variances are equal or not. To do so, we need to test homogeneity of variances between the
2 samples.
A basic data set is considered below, which will be used for testing variances as well as means.
This data set contains data collected from 2 samples. We wish to test if the means are statistically the same or do they differ.
Mean 4.633333 5.116666667 Means we fail to reject the fact that the
variances are statistically same.
F 1.091765
440
2 Sample T-Test: Solution
Hypothesis
Null Hypothesis – The means between the 2 samples are statistically the same
Alternate Hypothesis – The means between the 2 samples are statistically different.
Hypothesized Mean
Difference 0
df 10
t Stat -2.174864075
Normal
uous
n
onti ta
C Da
442
Unequal Variance Example
443
Hypothesis Testing Roadmap
Normal
uous
n
onti ta
C Da
444
Paired t-test
• A Paired t-test is used to compare the Means of two measurements from the same samples generally used as a before and after test.
• You can do Paired t test with Excel. This is appropriate for testing the difference between two Means when the data are paired and the paired
differences follow a Normal Distribution.
• Use the Paired t command to compute a confidence interval and perform a Hypothesis Test of the difference between population Means when
observations are paired. A paired t-procedure matches responses that are dependent or related in a pair-wise
manner. This matching allows you to account for
variability between the pairs usually resulting in
a smaller error term, thus increasing the sensitivity
of the Hypothesis Test or confidence interval.
– Ho: μδ = μo
– Ha: μδ ≠ μo
• Where μδ is the population Mean of the differences and μ0 is the hypothesized Mean of the differences, typically zero.
delta
()
Use Data Analysis in Excel
before after
445
Example
1. Practical Problem:
• We are interested in changing the sole material for a popular brand of shoes
for children.
• In order to account for variation in activity of children wearing the shoes, each
child will wear one shoe of each type of sole material. The sole material will
be randomly assigned to either the left or right shoe.
2. Statistical Problem:
H o: μ δ = 0
Ha: μδ ≠ 0
3. Paired t-test (comparing data that must remain paired).
α = 0.05 β = 0.10
We wish to test if the average dimensions for Mat B is higher than Mat A with Mat B given a
Mat A Mat B treatment post measurement of Mat A.
13 14
13.2 14.5 Null Hypothesis: Mean of Mat B and Mean of Mat A are statistically the same.
Alternate Hypothesis: Mean of Mat B and Mean of Mat A are statistically different.
13.4 14.4
13.5 13 Use Data Analysis in Excel.
14 14.5
13.8 13.5
13.9 13.2 t-Test: Paired Two Sample for Means
14 14
14.2 13
Variable 1 Variable 2
This means the treatment that was done to Mat A has not worked to its best.
t Stat -0.42075
You can use Paired t test as an option to validate improvements post solutions have been implemented in a project.
P(T<=t) one-tail 0.342507
Exercise objective: Utilize what you have learned to conduct and analyze a
paired t-test using MINITABTM.
2. The US quality manager wants to know with 95% confidence what the average
difference is between the lab located in Texas and the lab located in Mexico when
measuring Relative Crease Strength.
448
Paired t-test Exercise: Solution
Calc>Calculator
449
Paired t-test Exercise: Solution
We must confirm the differences (now in a new calculated column) are from a Normal
Distribution. This was confirmed with the Anderson-Darling Normality Test by doing a
graphical summary under Basic Statistics.
Summary
Summary for
for TX_MX-Diff
TX_MX-Diff
AAnderson-D
nderson-Darling
arlingNNormality
ormality Test
Test
AA-S quared
-S quared 0.45
0.45
PP-V
-Value
alue 0.222
0.222
MMean
ean 0.22727
0.22727
SStD
tDev
ev 0.37971
0.37971
VVariance
ariance 0.14418
0.14418
SSkew
kewness
ness -0.833133
-0.833133
Kurtosis
K urtosis -0.233638
-0.233638
NN 11
11
MMinimum
inimum -0.50000
-0.50000
1st
1stQQuartile
uartile -0.10000
-0.10000
MMedian
edian 0.40000
0.40000
3rd
3rdQQuartile
uartile 0.50000
0.50000
-0.50 -0.25 0.00 0.25 0.50 0.75 MMaximum 0.70000
-0.50 -0.25 0.00 0.25 0.50 0.75 aximum 0.70000
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forMMean
ean
-0.02782
-0.02782 0.48237
0.48237
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forMMedian
edian
-0.11644
-0.11644 0.50822
0.50822
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forSStD
tDev
95%
95%Confidence
ConfidenceIntervals
Intervals ev
0.26531
0.26531 0.66637
0.66637
Mean
Mean
Median
Median
0.0
0.0 0.2
0.2 0.4
0.4 0.6
0.6
450
Paired t-test Exercise: Solution
451
Paired t-test Exercise: Solution
Even though the Mean difference is 0.23, we have a 95% confidence interval that includes zero so we
know the 1-sample t-test’s null hypothesis was “failed to be rejected”. We cannot conclude the two
labs have a difference in lab results.
Histogram
Histogramof
of TX_MX-Diff
TX_MX-Diff
(with
(withHo
Hoand
and95%
95%t-confidence
t-confidence interval
interval for
for the
the mean)
mean)
55
44
33
Frequency
Frequency
452
Hypothesis Testing Roadmap
Normal
uous
n
onti ta
C Da
453
Summary
454
Hypothesis Testing Normal Data Part 2
Inferential Statistics
Hypothesis Testing ND P1
Calculate Sample Size
Analyze Results
Hypothesis Testing NND P1
455
Tests of Variance
Tests of Variance are used for both Normal and Non-normal Data.
Normal Data
– 1 Sample to a target
– 2 Samples – F-Test
– 3 or More Samples Bartlett’s Test
Non-Normal Data
– 2 or more samples Levene’s Test
456
1-Sample Variance
If the target variance lies inside the confidence interval, fail to reject the null
hypothesis.
– Ho: σ2Sample = σ2Target
– Ha: σ2Sample ≠ σ2Target
Use the sample size calculations for a 1 sample t-test since they are rarely
performed without performing a 1 sample t-test as well.
457
1-Sample Variance
1. Practical Problem:
• We are considering changing supplies for a part that we currently purchase
from a supplier that charges a premium for the hardening process and has a
large variance in their process.
• The proposed new supplier has provided us with a sample of their product.
They have stated they can maintain a variance of 0.10.
2. Statistical Problem:
Ho: σ2 = 0.10 or Ho: σ = 0.31
Ha: σ2 ≠ 0.10 Ha: σ ≠ 0.31
3. 1-sample variance:
α = 0.05 β = 0.10
458
1-Sample Variance
4. Sample Size:
• Open the MINITABTM worksheet: “Exh_Stat.MTW”
• This is the same file used for the 1 Sample t example.
– We will assume the sample size is adequate.
459
1-Sample Variance
Summary
Summary for
for Values
Values
AAnderson-D
nderson-Darling
arlingNNormality
ormality Test
Test
AA-S quared
-S quared 0.33
0.33
PP-V
-Value
alue 0.442
0.442
MMean
ean 4.7889
4.7889
SStD
tDev
ev 0.2472
0.2472
VVariance
ariance 0.0611
0.0611
SSkew
kewness
ness -0.02863
-0.02863
Kurtosis
Kurtosis -1.24215
-1.24215
NN 99
MMinimum
inimum 4.4000
4.4000
1st
1stQQuartile
uartile 4.6000
4.6000
MMedian
edian 4.7000
4.7000
3rd
3rdQQuartile
uartile 5.0500
5.0500
4.4
4.4 4.6
4.6 4.8
4.8 5.0
5.0 MMaximum 5.1000
aximum 5.1000
95%
95% CConfidence
onfidence Interv
Interval
alfor
forMMean
ean
4.5989
4.5989 4.9789
4.9789
95%
95% CConfidence
onfidence Interv
Interval
alfor
forMMedian
edian
4.6000
4.6000 5.0772
5.0772
95%
95% CConfidence
onfidence Interv
Interval
alfor
forSStDev
tDev
95%
95%Confidence
ConfidenceIntervals
Intervals
0.1670
0.1670 0.4736
0.4736
Mean
Mean
Median
Median
4.6
4.6 4.7
4.7 4.8
4.8 4.9
4.9 5.0
5.0 5.1
5.1
460
Test of Variance Example
1. Practical Problem:
We want to determine the effect of two different storage methods on the rotting
of potatoes. You study conditions conducive to potato rot by injecting potatoes
with bacteria that cause rotting and subjecting them to different temperature and
oxygen regimes. We can test the data to determine if there is a difference in the
Standard Deviation of the rot time between the two different methods.
2. Statistical Problem:
Ho: σ1 = σ2
Ha: σ1 ≠ σ2
461
Test of Variance Example
4. Sample Size:
α = 0.05 β = 0.10
EXH_AOV.MTW
462
Normality Test – Follow the Roadmap
5. Statistical Solution:
463
Normality Test – Follow the Roadmap
Probability
Probability Plot
Plot of
of Rot
Rot 11
Normal
Normal
99.9
99.9
Mean
Mean 4.871
4.871
StDev
StDev 0.9670
0.9670
99
99 NN 100
100
AADD 0.306
0.306
95
95 P-Value 0.559
P-Value 0.559
90
90
80
80
70
70
Percent
Percent
60
60
50
50
40
40
30
30
20
20
10
10
55
11
0.1
0.1
22 33 44 55 66 77 88
Rot
Rot 11
464
Test of Equal Variance
465
Test of Equal Variance
6. Practical Solution:
The difference between the Standard Deviations from the two samples is not
significant.
Test
Test for
for Equal
Equal Variances
Variances for
for Rot
Rot 11
F-Test
F-Test
Test Statistic
Test Statistic 0.74
0.74
Factors 11 P-Value
P-Value 0.298
0.298
Lev
Levene's
ene'sTest
Factors
Test
Test
TestStatistic 0.53
Use F-Test for 2 samples Statistic
P-Value
0.53
0.469
P-Value 0.469
Normally distributed data.
22
11
Factors
Factors
22
22 33 44 55 66 77
Rot
Rot11
466
Normality Test
Probability
Probability Plot
Plot of
of Rot
Rot
Normal
Normal
99
99
Mean
Mean 13.78
13.78
StDev
StDev 7.712
7.712
95
95 NN 18
18
AADD 0.285
0.285
90
90 P-Value
P-Value 0.586
0.586
80
80
70
70
Percent
Percent
60
60
50
50
40
40
30
30
The P-value is > 0.05
20
20 We can assume our data is
10
Normally Distributed.
10
55
11
-5
-5 00 55 10
10 15
15 20
20 25
25 30
30 35
35
Rot
Rot
467
Test for Equal Variance (Normal Data)
468
Test of Equal Variance
Test
Test for
for Equal
Equal Variances
Variances for
for Rot
Rot
F-Test
F-Test
Test Statistic
Test Statistic 0.68
0.68
10
10 P-Value
P-Value 0.598
0.598
Lev
Levene's
ene'sTest
Test
Temp
Temp
Test
TestStatistic
Statistic 0.05
0.05
P-Value
P-Value 0.824
0.824
16
16
22 44 66 88 10
10 12
12
95% Bonferroni Confidence Intervals for StDevs
95% Bonferroni Confidence Intervals for StDevs
Ho: σ11 = σ22
Ha: σ1≠ σ22
10
10
P-value > 0.05; there is no statistically
Temp
Temp
significant difference.
16
16
00 55 10
10 15
15 20
20 25
25
Rot
Rot
469
Test of Equal Variance
470
Continuous Data - Normal
471
Test For Equal Variances
472
Test For Equal Variances Graphical Analysis
Test
Test for
for Equal
Equal Variances
Variances for
for Rot
Rot
Temp
Temp Oxygen
Oxygen
Bartlett's
Bartlett'sTest
Test
22 Test Statistic
Test Statistic 2.71
2.71
P-Value
P-Value 0.744
0.744
10 66 Lev
Levene's
ene'sTest
Test
10
Test
TestStatistic
Statistic 0.37
0.37
P-Value
P-Value 0.858
0.858
10
10
22
16
16 66
10
10
00 20
20 40
40 60
60 80
80 100
100 120 120 140
140
95%
95%Bonferroni
BonferroniConfidence
Confidence Intervals
Intervals for
for StDevs
StDevs
473
Test For Equal Variances Statistical Analysis
474
Tests for Variance Exercise
Exercise objective: Utilize what you have learned to conduct and analyze a
test for equal variance using MINITABTM.
1. The quality manager was challenged by the plant director as to why the VOC
levels in the product varied so much. After using a Process Map, some potential
sources of variation were identified. These sources included operating shifts and
raw material supplier. Of course, the quality manager has already clarified the
Gage R&R results were less than 17% study variation so the gage was acceptable.
2. The quality manager decided to investigate the effect of the raw material supplier.
He wants to see if the variation of the product quality is different when using
supplier A than supplier B. He wants to be 95% confident the variances are
similar when using the two suppliers.
3. Use data ppm VOC and RM Supplier to determine if there is a difference between
suppliers.
475
Tests for Variance Exercise: Solution
First we want to do a graphical summary of the two samples from the 2 suppliers.
476
Tests for Variance Exercise: Solution
477
Tests for Variance Exercise: Solution
The P-value is greater than 0.05 for both Anderson-Darling Normality Tests so we
conclude the samples are from Normally Distributed populations because we “failed to
reject” the null hypothesis that the data sets are from Normal Distributions.
Median Median
478
Tests for Variance Exercise: Solution
479
Tests for Variance Exercise: Solution
480
Tests for Variance Exercise: Solution
Because the 2 populations were considered to be Normally Distributed, the F-test is used to evaluate
whether the variances (Standard Deviation squared) are equal.
The P-value of the F-test was greater than 0.05 so we “fail to reject” the null hypothesis.
So once again in English: The variances are equal between the results from the two suppliers on our
product’s ppm VOC level.
Test
Testfor
for Equal
Equal Variances
Variances for
for ppm
ppmVOC
VOC
F-Test
F-Test
Test
TestStatistic
Statistic 1.16
1.16
AA P-Value
P-Value 0.806
0.806
Supplier
RMSupplier
Lev
Levene's
ene'sTest
Test
Test
TestStatistic
Statistic 0.02
0.02
P-Value
P-Value 0.890
0.890
RM
BB
44 66 88 10
10 12
12 14
14
95% Bonferroni Confidence Intervals for StDevs
95% Bonferroni Confidence Intervals for StDevs
AA
Supplier
RMSupplier
RM
BB
20
20 25
25 30
30 35
35 40
40 45
45 50
50
ppm
ppmVOC
VOC
481
Hypothesis Testing Roadmap
Normal
uous
n
onti ta
C Da
482
Purpose of ANOVA
Analysis of variance extends the two sample t-test for testing the equality of two
population Means to a more general null hypothesis of comparing the equality
of more than two Means, versus them not all being equal.
– The classification variable, or factor, usually has three or more levels (If
there are only two levels, a t-test can be used).
– Allows you to examine differences among means using multiple
comparisons.
– The ANOVA test statistic is:
(δ)
X
X
X X
X
X X X
μ1 μ2
484
Calculating ANOVA
Where:
G - the number of groups (levels in the study)
xij = the individual in the jth group
nj = the number of individuals in the jth group or level
X = the grand Mean
Xj = the Mean of the jth group or level
delta
(δ) Within Group Variation
j1
nj (Xj X )
2
(X ij X)
2
(X
j1 i 1
ij X) 2
j1 i 1
485
Alpha Risk and Pair-Wise t-tests
The alpha risk increases as the number of Means increases with a pair-wise t-
test scheme. The formula for testing more than one pair of Means using a t-test
is:
1 1 α
k
486
Three Samples
We have three potential suppliers that claim to have equal levels of quality.
Supplier B provides a considerably lower purchase price than either of the
other two vendors. We would like to choose the lowest cost supplier but we
must ensure that we do not effect the quality of our raw material.
487
Follow the Roadmap…Test for Normality and Equal Variances
All three suppliers samples are Box Plots don’t reveal unusual trends
Normally Distributed.
SUMMARY
488
Within Groups 1.6412 12 0.136767
Sample Size
489
ANOVA Assumptions
Stat>ANOVA>One-Way Unstacked>Graphs
491
Histogram of Residuals
Histogram
Histogramof
of the
the Residuals
Residuals
(responses
(responses are
are Supplier
Supplier A,
A, Supplier
Supplier B,
B, Supplier
Supplier C)
C)
55
44
Frequency
Frequency
33
22
11
00
-0.6
-0.6 -0.4
-0.4 -0.2
-0.2 0.0
0.0 0.2
0.2 0.4
0.4 0.6
0.6
Residual
Residual
492
Normal Probability Plot of Residuals
Normal
Normal Probability
Probability Plot
Plot of
of the
the Residuals
Residuals
(responses
(responses are
are Supplier
Supplier A,
A, Supplier
Supplier B,
B, Supplier
Supplier C)
C)
99
99
95
95
90
90
80
80
70
70
Percent
Percent
60
60
50
50
40
40
30
30
20
20
10
10
55
11
-1.0
-1.0 -0.5
-0.5 0.0
0.0 0.5
0.5 1.0
1.0
Residual
Residual
493
Residuals versus Fitted Values
Residuals
Residuals Versus
Versus the
theFitted
FittedValues
Values
(responses
(responses are
are Supplier
Supplier A,
A, Supplier
Supplier B,
B,Supplier
Supplier C)
C)
0.75
0.75
0.50
0.50
0.25
0.25
Residual
Residual
0.00
0.00
-0.25
-0.25
-0.50
-0.50
3.65
3.65 3.70
3.70 3.75
3.75 3.80
3.80 3.85
3.85 3.90
3.90 3.95
3.95 4.00
4.00 4.05
4.05
Fitted Value
Fitted Value
494
ANOVA Exercise
1. The quality manager was challenged by the plant director as to why the
VOC levels in the product varied so much. The quality manager now
wants to find if the product quality is different because of how the shifts
work with the product.
2. The quality manager wants to know if the average is different for the
ppm VOC of the product among the production shifts.
495
ANOVA Exercise: Solution
496
ANOVA Exercise: Solution
497
ANOVA Exercise: Solution
Summary
Summaryfor
forppm
ppmVOC
The P-value is greater than 0.05 for both Shift
VOC
Shift==11
P-Value 0.446
Median
Median
35 40 45 50
35 40 45 50
Summary Summary
Summaryfor
forppm
ppmVOC P-Value 0.658
Summaryfor
forppm
ppmVOC
VOC P-Value 0.334 VOC
Shift
Shift
Shift==22 Shift==33
A nderson-D arling N ormality Test A nderson-D arling N ormality Test
A nderson-D arling N ormality Test A nderson-D arling N ormality Test
A -S quared 0.37 A -S quared 0.24
A -S quared 0.37 A -S quared 0.24
P -V alue 0.334 P -V alue 0.658
P -V alue 0.334 P -V alue 0.658
M ean 34.625 M ean 28.000
M ean 34.625 M ean 28.000
S tD ev 5.041 S tD ev 6.525
S tD ev 5.041 S tD ev 6.525
V ariance 25.411 V ariance 42.571
V ariance 25.411 V ariance 42.571
S kew ness -0.74123 S kew ness 0.06172
S kew ness -0.74123 S kew ness 0.06172
Kurtosis 1.37039 Kurtosis -1.10012
Kurtosis 1.37039 Kurtosis -1.10012
N 8 N 8
N 8 N 8
M inimum 25.000 M inimum 19.000
M inimum 25.000 M inimum 19.000
1st Q uartile 31.750 1st Q uartile 22.000
1st Q uartile 31.750 1st Q uartile 22.000
M edian 35.500 M edian 28.000
M edian 35.500 M edian 28.000
20 25 30 35 40 45 50 3rd Q uartile 37.000 20 25 30 35 40 45 50 3rd Q uartile 32.750
20 25 30 35 40 45 50 3rd Q uartile 37.000 20 25 30 35 40 45 50 3rd Q uartile 32.750
M aximum 42.000 M aximum 38.000
M aximum 42.000 M aximum 38.000
95% C onfidence Interv al for M ean 95% C onfidence Interv al for M ean
95% C onfidence Interv al for M ean 95% C onfidence Interv al for M ean
30.411 38.839 22.545 33.455
30.411 38.839 22.545 33.455
95% C onfidence Interv al for M edian 95% C onfidence Interv al for M edian
95% C onfidence Interv al for M edian 95% C onfidence Interv al for M edian
30.614 37.322 20.871 33.322
95% Confidence Intervals 30.614 37.322 95% Confidence Intervals 20.871 33.322
95% Confidence Intervals 95% C onfidence Interv al for S tD ev 95% Confidence Intervals 95% C onfidence Interv al for S tD ev
95% C onfidence Interv al for S tD ev 95% C onfidence Interv al for S tD ev
Mean Mean 4.314 13.279
Mean
3.333 10.260 Mean 4.314 13.279
3.333 10.260
Median Median
Median Median
498
ANOVA Exercise: Solution
499
ANOVA Exercise: Solution
The P-value of the F-test was greater than 0.05 so we “fail to reject” the null
hypothesis.
Test
Testfor
for Equal
Equal Variances
Variances for
for ppm
ppmVOC
VOC
Bartlett's
Bartlett'sTest
Test
Test
TestStatistic
Statistic 0.63
0.63
11 P-Value
P-Value 0.729
0.729
Lev
Levene's
ene'sTest
Test
Test
TestStatistic
Statistic 0.85
0.85
P-Value
P-Value 0.440
0.440
Shift
Shift
22
33
22 44 66 88 10
10 12
12 14
14 16
16 18
18
95% Bonferroni Confidence Intervals for StDevs
95% Bonferroni Confidence Intervals for StDevs
We must look at the Residual Plots to be sure our ANOVA analysis is valid.
Since our residuals look Normally Distributed and randomly patterned, we will
assume our analysis is correct.
Residual
Residual Plots
Plots for
for ppm
ppmVOC
VOC
Normal
NormalProbability
Probability Plot
Plot Residuals
ResidualsVersus
Versusthe
theFitted
Fitted Values
Values
99
99 NN 24
24 10
AD 0.255
10
90 AD 0.255
90 P-Value 0.698
P-Value 0.698 55
Residual
Percent
Residual
Percent
50
50 00
10 -5
-5
10
11 -10
-10
-10
-10 00 10
10 30
30 35
35 40
40
Residual
Residual Fitted
FittedValue
Value
Histogram
Histogramof
of the
theResiduals
Residuals Residuals
ResidualsVersus
Versusthe
theOrder
Orderof
of the
theData
Data
4.8 10
10
4.8
3.6 55
Frequency
3.6
Frequency
Residual
Residual
2.4 00
2.4
1.2 -5
-5
1.2
0.0 -10
-10
0.0
-10
-10 -5
-5 00 55 10
10 22 44 66 88 10
10 12
12 14
14 16
16 18
18 20
20 22
22 24
24
Residual
Residual Observation Order
Observation Order
502
ANOVA Exercise: Solution
Since the P-value of the ANOVA test is less than 0.05, we “reject” the null
hypothesis that the Mean product quality as measured in ppm VOC is the
same from all shifts.
We “accept” the alternate hypothesis that the Mean product quality is
different from at least one shift.
503
Summary
504
Hypothesis Testing Non Normal Data Part 1
Inferential Statistics
Hypothesis Testing ND P1
Hypothesis Testing ND P2
Equal Variance Tests
Hypothesis Testing NND P1
Tests for Medians
Hypothesis Testing NND P2
505
Non-Normal Hypothesis Tests
At this point we have covered the tests for determining significance for Normal
Data. We will continue to follow the roadmap to complete the test for Non-
normal Data with Continuous Data.
Later in the module we will use another roadmap that was designed for Discrete
Data.
– Recall that Discrete Data does not follow a Normal Distribution, but
because it is not Continuous Data, there are a separate set of tests to
properly analyze the data.
We will focus on Skewness for the remaining tests for Continuous Data.
507
Hypothesis Testing Roadmap
uous
nt in
Co Data Non Normal
508
Test of Equal Variance
– Ho: σ1 = σ2 = σ3 …
– Ha: At least one is different.
509
Follow the Roadmap…
0.1
EXH_AOV.MTW -5.0 -2.5 0.0 2.5 5.0 7.5 10.0
Rot 2
510
Test of Equal Variance Non-Normal Distribution
Test
Testfor
for Equal
Equal Variances
Variances for
for Rot
Rot22
F-Test
F-Test
Test Statistic
Test Statistic 1.75
1.75
11 P-Value
P-Value 0.053
0.053
Factors2
Lev
Levene's
ene'sTest
Factors2
Test
Test
TestStatistic
Statistic 0.03
0.03
P-Value
P-Value 0.860
0.860
22
1.0
1.0 1.2
1.2 1.4
1.4 1.6
1.6 1.8
1.8 2.0
2.0 2.2
2.2
95% Bonferroni Confidence Intervals for StDevs
95% Bonferroni Confidence Intervals for StDevs
11
Factors2
Factors2
22
00 22 44 66 88 10
10
Rot
Rot22
511
Making Conclusions
When testing >2 samples with Normal Distribution, use Bartlett’s test:
– To determine whether multiple Normal Distributions have equal variance.
When testing 2 or more samples with Non-normal Distributions, use Levene’s test:
– To determine whether two or more distributions have Equal Variance.
Our focus for this module when working with Non-normal Distributions.
512
Exercise
A credit card company wants to understand the need for customer service
personnel. The company thinks there is variability impacting the efficiency
of its customer service staff. The credit card company has two types of
cards. The company wants to see if there is more variability in one type of
customer card than another. The Black Belt was selected and told to give
with 95% confidence the answer of similar variability between the two card
types.
HYPOTESTSTUD.MPJ
513
Test for Equal Variance Example: Solution
514
Test for Equal Variance Example: Solution
515
Test for Equal Variance Example: Solution
516
Test for Equal Variance Example: Solution
Data>Stack>Columns…
517
Test for Equal Variance Example: Solution
518
Nonparametric Tests
~ ~ ~
Target X X1 X2
519
Mean and Median
In general, nonparametric tests do the following: rank order the data, sum the data by ranks, sign the data
above or below the target, and calculate, compare and test the Median.
Comparisons and tests about the Median make nonparametric tests useful with very Non-normal Data.
Note: This Graphical Summary provides the confidence interval for the Median.
Mean
Median
520
MINITABTM’s Nonparametrics
1-Sample Sign: performs a one-sample sign test of the Median and calculates the
corresponding point estimate and confidence interval. Use this test as an alternative to
one-sample Z and one-sample t-tests.
1-Sample Wilcoxon: performs a one-sample Wilcoxon signed rank test of the Median and
calculates the corresponding point estimate and confidence interval (more
discriminating or efficient than the sign test). Use this test as a nonparametric
alternative to one-sample Z and one-sample t-tests.
Mann-Whitney: performs a Hypothesis Test of the equality of two population Medians and
calculates the corresponding point estimate and confidence interval. Use this test as a
nonparametric alternative to the two-sample t-test.
Kruskal-Wallis: performs a Hypothesis Test of the equality of population Medians for a
one-way design. This test is more powerful than Mood’s Median (the confidence
interval is narrower, on average) for analyzing data from many populations, but is less
robust to outliers. Use this test as an alternative to the one-way ANOVA.
Mood’s Median Test: performs a Hypothesis Test of the equality of population Medians in
a one-way design. Test is similar to the Kruskal-Wallis Test. Also referred to as the
Median test or sign scores test. Use as an alternative to the one-way ANOVA.
521
1-Sample Sign Test
This test is used when you want to compare the Median of one distribution to a
target value.
– Must have at least one column of numeric data. If there is more than one column
of data, MINITABTM performs a one-sample Wilcoxon test separately for each
column.
The hypotheses:
– H0: M = Mtarget
– Ha: M ≠ Mtarget
Interpretation of the resulting P-value is the same.
Note: For the purpose of calculating sample size for a non-parametric (Median)
test use:
n t test
n non -parametric
0.864
522
1-Sample Example
1. Practical Problem:
Our facility requires a cycle time from an improved process of 63 minutes. This process supports
the customer service division and has become a bottleneck to completion of order processing. To
alleviate the bottleneck the improved process must perform at least at the expected 63 minutes.
2. Statistical Problem:
Ho: M = 63
Ha: M ≠ 63
4. Sample Size:
This data set has 500 samples (well in excess of necessary sample size).
523
1-Sample Example
=
Sign Test for Median: Pos Skew
Sign Test of Median = 63.00 versus = 63.00
N Below Equal Above P Median
Pos Skew 500 37 0 463 0.0000 65.70
524
1-Sample Example
525
1-Sample Example
526
1-Sample Example
527
1 Sample Example: Solution
HYPOTESTSTUD.MPJ
528
1 Sample Example: Solution
According to the hypothesis the Mine Manager feels he is achieving his target of 2.1 tons/day.
H0: M = 2.1 tons/day Ha: M ≠ 2.1 tons/day
Since we are using one sample, we have a choice of choosing either a 1 Sample-Sign or 1 Sample Wilcoxon.
For this example we will use a 1 Sample-Sign.
Stat>Nonparametrics>1-Sample Sign
529
1 Sample Example: Solution
The Black Belt in this case does not agree; based on this data the Mine
Manager is not achieving his target of 2.1 tons/day.
We disagree!
530
Mann-Whitney Example
The Mann-Whitney test is used to test if the Medians for 2 samples are different.
2. Ho: M1 = M2
Ha: M1 ≠ M2
4. There are 200 data points for each machine, well over the minimum sample
necessary.
531
Mann-Whitney Example
5. Statistical Conclusion
Probability
Probability Plot
Plotof
of Mach
MachAA
Normal
Normal
99.9
99.9
Mean
Mean 15.24
15.24
StDev
StDev 5.379
5.379
99
99 NN 200
200
AADD 1.550
1.550
95
95 P-Value
P-Value <0.005
<0.005
90
90
80
80
70
Probability
Probability Plot
Plotof
of Mach
MachBB
Percent
70
Normal
Percent
60
60 Normal
50
50
40 99.9
99.9
40 Mean 16.73
30
30 Mean 16.73
20 StDev
StDev 5.284
5.284
20 99
99 NN 200
200
10
10 AADD 0.630
0.630
55 95
95 P-Value
P-Value 0.099
0.099
90
90
11
80
80
70
Percent
70
Percent
0.1
0.1 60
60
00 10 20 50
50 30 40
10 20 40 30 40
40
Mach
MachAA 30
30
20
20
10
10
55
11
0.1
0.1
00 55 10
10 15
15 20
20 25
25 30
30 35
35
Mach B
Mach B
532
Mann-Whitney Example
Stat>Nonparametric>Mann-Whitney…
533
Exercise
HYPOTESTSTUD.MPJ
534
Mann-Whitney Example: Solution
535
Mann-Whitney Example: Solution
As you can see there is a difference in the Median between CallsperWk1 and
CallsperWk2.
Therefore, there is not a difference in call volume between the two different card
types.
536
Mood’s Median Test
= = ?
537
Follow the Roadmap…Normality
Summary
Summary for
for Recovery
Recovery
Location
Location==Savannah
Savannah
AAnderson-D
nderson-Darling
arlingNNormality
ormality Test
Test
AA-S
-Squared
quared 0.81
0.81
PP-V
-Value
alue 0.032
0.032
MMean
ean 87.660
87.660
SStD
tDev
ev 7.944
7.944
VVariance
ariance 63.113
63.113
SSkew
kewness
ness -0.15286
-0.15286
Kurtosis
Kurtosis -1.11764
-1.11764
NN 25
25
MMinimum
inimum 75.300
75.300
1st
1stQQuartile
uartile 79.000
79.000
MMedian
edian 87.500
87.500
78
78 84
84 90
90 96
96
3rd
3rdQQuartile
uartile 96.550
96.550
MMaximum
aximum 99.200
99.200
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forMMean
ean
84.381
84.381 90.939
90.939
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forMMedian
edian
86.179
86.179 90.080
90.080
9955%
% CConfidence
onfidenceInter
Intervals 95%
v als 95% CConfidence
onfidenceInterv
Intervalalfor
forSStDev
tDev
Mean 6.203
6.203 11.052
11.052
Mean
Median
Median
84.0
84.0 85.5
85.5 87.0
87.0 88.5
88.5 90.0
90.0 91.5
91.5
538
Follow the Roadmap…Normality
Summary
Summary for
for Recovery
Recovery
Location
Location==Bangor
Bangor
AAnderson-D
nderson-Darling
arlingNNormality
ormality Test
Test
AA-S
-Squared
quared 0.72
0.72
PP-V
-Value
alue 0.045
0.045
MMean
ean 93.042
93.042
SStDev
tDev 5.918
5.918
VVariance
ariance 35.017
35.017
SSkew
kewness
ness -1.81758
-1.81758
Kurtosis
Kurtosis 4.66838
4.66838
NN 13
13
MMinimum
inimum 76.630
76.630
1st
1stQQuartile
uartile 90.600
90.600
MMedian
3rd
edian Summary
Summary for
94.800
94.800
for Recovery
Recovery
78
78
84
84
90
90
96
96 3rdQQuartile
uartile 97.350
97.350
MMaximum
aximum 99.700 Location = Ankhar
99.700 Location = Ankhar
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forMMean AAnderson-D
ean nderson-Darling
arlingNNormality
ormality Test
Test
89.466 96.617 AA-S
89.466 96.617 -Squared
quared 0.86
0.86
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forMMedian
edian PP-V
-Value
alue 0.022
0.022
90.637
90.637 97.036
97.036 MMean 88.302
ean 88.302
9955%
% CConfidence
onfidenceInter
Intervals 95%
v als 95% CConfidence
onfidenceInterv
Intervalalfor
forSStDev
tDev SStD
tDev
ev 6.929
6.929
4.243 9.768 VVariance
ariance 48.008
48.008
Mean 4.243 9.768
Mean SSkew
kewness
ness -0.105610
-0.105610
Median
Kurtosis
Kurtosis 0.182123
0.182123
Median NN 20
20
90 92 94 96 98
90 92 94 96 98 MMinimum
inimum 73.500
73.500
1st
1stQQuartile
uartile 85.150
85.150
MMedian
edian 88.425
88.425
78
78
84
84
90
90
96
96
3rd
3rdQQuartile
uartile 89.700
89.700
MMaximum
aximum 99.450
99.450
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forMMean
ean
85.059
85.059 91.545
91.545
95%
95% CConfidence
onfidenceInterv
Intervalalfor
forMMedian
edian
86.735
86.735 89.299
89.299
9955%
% CConfidence
onfidenceInter
Intervals 95%
v als 95% CConfidence
onfidenceInterv
Intervalalfor
forSStD
tDev
ev
Mean 5.269
5.269 10.120
10.120
Mean
Median
Median
85
85 86
86 87
87 88
88 89
89 90
90 91
91
539
Follow the Roadmap…Equal Variance
Test
Testfor
for Equal
Equal Variances
Variances for
for Recovery
Recovery
Bartlett's
Bartlett'sTest
Test
Test
TestStatistic
Statistic 1.33
1.33
Ankhar
Ankhar P-Value
P-Value 0.514
0.514
Lev
Levene's
ene'sTest
Test
Test Statistic
Test Statistic 1.02
1.02
Location
Location P-Value
P-Value 0.367
0.367
Bangor
Bangor
Savannah
Savannah
33 44 55 66 77 88 99 10
10 11 11 1212
95%
95%Bonferroni
BonferroniConfidence
Confidence Intervals
Intervals for
for StDevs
StDevs
540
Mood’s Median Test
Statistical Solution: Since the P-value of the Mood’s Median test is less than
0.05, we reject the null hypothesis.
Practical Solution: Bangor has the highest recovery of all three facilities.
We observe the confidence intervals for the Medians of
the 3 populations. Note there is no overlap of the 95%
confidence levels for Bangor—so we visually know the
P-value is below 0.05.
541
Kruskal-Wallis Test
Using the same data set, analyze using the Kruskal-Wallis test.
H = 6.86 DF = 2 P = 0.032
H = 6.87 DF = 2 P = 0.032 (adjusted for ties)
This output is the “least friendly” to interpret. Look for the P-value which tells us
we reject the null hypothesis. We have the same conclusion as with the Mood’s
Median test.
542
Exercise
HYPOTESTSTUD.MPJ
543
Pagers Defect Rate Example: Solution
Instead of performing a
Normality Test, we can find
the P-value using the
Graphical Summary in
MINITABTM.
545
Pagers Defect Rate Example: Solution
546
Unequal Variance
547
Example
Model A and Model B are similar in nature (not exact), but are manufactured in
the same plant.
– Check for Normality:
Var_Comp.mtw
Percent
60 60
50 50
40 40
30 30
20 20
10 10
5 5
1 1
8.5 9.0 9.5 10.0 10.5 11.0 11.5 12.0 -5.0 -2.5 0.0 2.5 5.0 7.5 10.0
Model A Model B
548
Example
Test
TestStatistic
Statistic 4.47
4.47
P-Value
P-Value 0.049
0.049
Model
ModelBB
00 11 22 33 44 55 66 77
95% Bonferroni Confidence Intervals for StDevs
95% Bonferroni Confidence Intervals for StDevs
Model
ModelAA
idvar
idvar
Model
ModelBB
00 22 44 66 88 10
10 12
12
Data
Data
The P-value is just under the limit of .05. Whenever the result
is borderline, as in this case, use your process knowledge to
make a judgment.
549
Example
Median Median
Model A
Model B
-0.0 1.6 3.2 4.8 6.4 8.0 9.6 11.2
Data
550
Black Belt Aptitude Exercise
HYPOTESTSTUD.MPJ
551
Black Belt Aptitude Exercise: Solution
552
Black Belt Aptitude Exercise: Solution
553
Summary
554
Hypothesis Testing Non Normal Data Part 2
Inferential Statistics
Hypothesis Testing ND P1
Hypothesis Testing ND P2
555
Hypothesis Testing Roadmap Attribute Data
556
Sample Size and Types of Data
MINITABTM can estimate sample sizes, but remember the smaller the
difference that needs to be detected the larger the sample size will be!
557
Proportion versus a Target
This test is used to determine if the process proportion (p) equals some
desired value, p0.
The hypotheses:
– H0: p = p 0
– Ha: p ¹ p 0
Z
pˆ p
0
(normal approximation) p 1 p n
obs
0 0
558
Proportion versus a Target
4. Sample size:
559
Proportion versus a Target
Our sample included 500 shipped items of which 480 were accurate.
X 480
p̂ 0.96
n 500
560
Proportion versus a Target
561
Exercise
562
Proportion vs Target Example: Solution
563
Proportion vs Target Example: Solution
?
Now let us calculate if we receive
our bonus…
X 1680
p̂ 0.84
n 2000
564
Comparing Two Proportions
This test is used to determine if the process defect rate (or proportion, p) of one
sample differs by a certain amount D from that of another sample (e.g., before and
after your improvement actions)
The hypotheses:
H0: p1 - p2 = D
Ha: p1 - p2 ¹ D
Take a few moments to practice calculating the minimum sample size required to
detect a difference between two proportions using a power of 0.90.
For a more conservative estimate when the null hypothesis is close to 100, use smaller
proportion for p1. When the null hypothesis is close to 0, use the larger proportion for
p1.
a p1 p2 n
5% .01 0.79 0.8 ___________
5% .01 0.81 0.8 ___________
5% .02 0.08 0.1 ___________
5% .02 0.12 0.1 ___________
5% .01 0.47 0.5 ___________
5% .01 0.53 0.5 ___________
566
Proportion versus a Target
1. Shipping accuracy must improve from a historical baseline of 85% towards a target of
95%. Determine if the process improvements made have increased the accuracy.
2. Hypotheses:
– H0: p1 – p2 = 0.0 Stat>Power and Sample Size> 2 Proportions…
– Ha: p1 – p2 ¹ 0.0
3. Two sample proportion test
– Choose a = 5%
4. Sample size:
Sample Target
Proportion 1 Size Power Actual Power
0.85 188 0.9 0.901 451
The sample size is for each group.
567
Comparing Two Proportions
568
Comparing Two Proportions
569
Exercise
1. Who is worse?
2. Is the sample size large enough?
570
2 Proportion vs Target Example: Solution
X1 47
Boris p̂1 0.132
n1 356
X2 99
Igor p̂ 2 0.173
n 2 571
571
2 Proportion vs Target Example: Solution
572
Contingency Tables
This test generally works the best with 5 or more observations in each cell.
Observations can be pooled by combining cells.
573
Contingency Tables
The null hypothesis is that the population proportions of each group are the
same.
– H0: p1 = p2 = p3 = … = pn
– Ha: at least one p is different
observed expected
2
expected
574
Test Statistic Calculations
Chi-square Test
r c (O ij E ij ) 2
χ
2
o Where:
From the Chi-Square Table Ftotal = total frequency for the table
n = degrees of freedom [(r-1)(c-1)]
575
Contingency Table Example
1. Larry, Curley and Moe are order entry operators and you suspect that
one of them has a lower defect rate than the others.
2. Ho: pMoe = pLarry = pCurley
577
Contingency Table Example
Now use these proportions to calculate the expected frequencies in each cell:
0.306*45 = 13.8
0.694 * 38 = 26.4
578
Contingency Table Example
observed - expected 2
expected
579
Contingency Table Example
580
Contingency Table Example
Critical Value:
• Like any other Hypothesis Test, compare the observed statistic with the critical
statistic. We decide a = 0.05, what else do we need to know?
• For a chi-square distribution, we need to specify n, in a Contingency Table:
n = (r - 1)(c - 1), where
r = # of rows
c = # of columns
• In our example, we have 2 rows and 3 columns, so n = 2
• What is the critical chi-square? For a Contingency Table, all the risk is in the
right hand tail (i.e. a one-tail test); look it up in MINITABTM using
Calc>Probability Distributions>Chisquare…
2 5.99
χ crit
581
Contingency Table Example
Graphical Summary:
Since the observed chi-square exceeds the critical chi-square, we reject the
null hypothesis that the defect rate is independent of which person enters the
orders.
0.5
0.4
0.3
Accept Reject
f
0.2 obs
2 7.02
0.1
0.0
0 1 2 3 4 5 6 7 8
chi-square crit 5.99
2
582
Contingency Table Example
Using MINITABTM
• Notice that row labels are not necessary and row and column totals are not
used, just the observed counts for each cell.
583
Contingency Table Example
2 20 30 25 75
17.36 26.39 31.25
0.401 0.494 1.250
Total 25 38 45 108
584
Exercise
• You are the quotations manager and your team thinks that the reason
you don’t get a contract depends on its complexity.
• You determine a way to measure complexity and classify lost contracts
as follows:
585
Contingency Table Example: Solution
Stat>Tables>Chi-Square Test
586
Contingency Table Example: Solution
587
Overview
588
Summary
589
Analyze
Wrap Up and Action Items
Analyze Wrap Up Overview
• Locate the variables which are significantly impacting your Primary Metric. Then
establish Root Causes for “X” variables using Inferential Statistical Analysis such
as Hypothesis Testing and Simple Modeling.
591
Six Sigma Behaviors
• Embracing change
• Continuous learning
• Being rigorous
Each
Each“player”
“player”ininthe
theSix
SixSigma
Sigmaprocess
processmust
mustbe
be
AAROLE
ROLEMODEL
MODEL
for
forthe
theSix
SixSigma
Sigmaculture.
culture.
592
Analyze Deliverables
• Listed below are the Analyze Phase deliverables that each candidate will present
in a Power Point presentation at the beginning of the Control Phase training.
• At this point you should all understand what is necessary to provide these
deliverables in your presentation.
– Team Members (Team Meeting Attendance)
– Primary Metric
– Secondary Metric(s)
– Data Demographics
– Hypothesis Testing (applicable tools)
– Modeling (applicable tools)
– Strategy to reduce X’s
– Project Plan
– Issues and Barriers It’s your show!
593
Analyze Phase - The Roadblocks
Look for the potential roadblocks and plan to address them before they
become problems:
– Lack of data
– Data presented is the best guess by functional managers
– Team members do not have the time to collect data
– Process participants do not participate in the analysis planning
– Lack of access to the process
594
DMAIC Roadmap
Process Owner
Champion/
Estimate COPQ
Establish Team
Measure
595
Analyze Phase
Collect Data
Statistically
Significant?
N
Y
Update FMEA
N
Practically
Significant?
Root
Cause
N
Y
Identify Root Cause
596
Analyze Phase Checklist
Analyze Questions
General Questions
• Are there any issues or barriers that prevent you from completing this phase?
• Do you have adequate resources to complete the project?
597
Planning for Action
599
Improve Phase
Process Modeling
Welcome to Improve
Correlation
Designing Experiments
Experimental Methods
601
Correlation
• The primary purpose of linear correlation analysis is to measure the strength of linear
association between two variables (X and Y).
• If X increases and there is no definite shift in the values of Y, there is no correlation or no
association between X and Y.
• If X increases and there is a shift in the values of Y, there is a correlation.
• The correlation is positive when Y tends to increase and negative when Y tends to decrease.
• If the ordered pairs (X, Y) tend to follow a straight line path, there is a linear correlation.
• The preciseness of the shift in Y as X increases determines the strength of the linear
correlation.
• To conduct a linear correlation analysis you need:
– Bivariate Data – Two pieces of data that are variable
– Bivariate data is comprised of ordered pairs (X/Y)
– X is the independent variable
– Y is the dependent variable
602
Correlation Coefficient
603
Types and Magnitude of Correlation
110 110 85
100
100
90
90
75
80
Output
Output
Output
80
70
70
60 65
60
50
50
40
55
40
30
Output
70
70
60
65
60
50
50
40
40 55
30
0 10 20 30 40 50 60 70 80 0 10 20 30 40 50 10 20 30 40 50 60
604
Limitations of Correlation
• A strong positive or negative correlation between X and Y does not indicate causality.
• Correlation provides an indication of the strength but does not provide us with an exact
numerical relationship (i.e. Y=f(x)).
• The magnitude of the Correlation Coefficient is somewhat relative and should be used with
caution.
• Just like any other statistic, you need to assess whether the Correlation Coefficient is
statistically significant, as well as practically significant.
• As usual, statistical significance is judged by comparing a P-value with the chosen degree
of alpha risk.
• Guidelines for practical significance are as follows:
– If | r | > 0.80, relationship is practically significant
– If | r | < 0.20, relationship is not practically significant
Areaofofnegative
Area negative Area of positive
linearcorrelation
correlation No linear correlation linear correlation
linear
RB Stats Correlation.mtw
X values Y values
The Correlation Coefficient [r]: Payton carries Payton yards
• Is a positive value if one variable increases 196 679
as the other variable increases. 311 1390
• Is a negative value if one variable 339 1852
decreases as the other increases. 333 1359
369 1610
317 1460
339 1222
148 596
Correlation Formula
314 1421
381 1684
( X i X )(Yi Y )
r 324 1551
( X i X ) 2 (Yi Y ) 2 321 1333
146 586
606
Correlation Analysis
1. With the data set, first you need to create a scatter plot. The scatter plot will give you a visual aid to how this model appears.
2. Then, you need to determine the Pearson’s Coefficient of Correlation. If the Pearson’s Coefficient of Correlation is > 0.8 or < -0.8, proceed to Step
3.
4. Use the p-values to decide which terms to include in the final regression equation.
6. Check if the difference between the Predicted values and Actual values (Known as Residuals), are normally distributed.
7. Non-normal residuals would often mean outliers in the residuals, which could be due to unusual observations resulting from special causes of
variations.
607
Correlation Analysis
Step 1 – The Pearson’s Coefficient of Correlation is 0.93. Check the Simple Regression sheet to know how the Pearson’s Coefficient of
Correlation has been arrived at.
Step 2 – As Pearson’s Coefficient of Correlation is 0.93, i.e. > 0.8, a strong correlation between Y and X is established.
Step 3 – Study the R-Square value. R-Square value is 0.8733, i.e. 87.33%. R-Square > 64%.
Step 4 – Now to study the p-values to statistically determine which of the components of the regression equation y = 4.916x – 163.5.
Step 5 – If p-values of components > 0.05, the component needs to be omitted. If p-values of components < 0.05, the component needs to be
included in the model.
608
Correlation Analysis
SUMMARY OUTPUT
Regression Statistics
Multiple R 0.934531611
R Square 0.873349332
Adjusted R
Square 0.861835635
Observations 13
ANOVA
df SS MS F Significance F
Regression 1
Step 6 – The p-value of the Intercept indicates 1798586.901 1798587
a non-significant p-value. 75.8530755
That means 2.88668E-06
the constant would need to be omitted from the model. The X Variable
shows a significant p-value, that means the final regression equation should be
Residual 11 260826.0217 23711.46
Y = 4.916x.
Total 12 2059412.923
Step 7 – Let us now plot the predicted values for the same value of x, and find out Residuals.
Lower Upper
Coefficients Standard Error t Stat P-value Lower 95% Upper 95% 95.0% 95.0%
The normal probability plot and the residuals plot do not indicate the presence of any special patterns.
Statistical Solution
3.The new value of Payton Yards will be the statistical solution desired from the project.
610
Modeling Y=f(x) Exercise: Question 3 Solution
If Dorsett carries the football 325 times the predicted value would be
determined as follows…
611
Summary
612
Advanced Modeling & Regression
Review Corr./Regression
Process Modeling: Regression
Non-Linear Regression
Advanced Process Modeling: MLR
Transforming Process Data
Designing Experiments
Multiple Regression
Experimental Methods
613
Correlation and Linear Regression Review
614
Correlation Review
Correlation is used to measure the linear relationship between two Continuous Variables
(bi-variate data).
Pearson Correlation Coefficient “r” will always fall between –1 and +1.
A Correlation of –1 indicates a strong negative relationship, one factor increases the
other decreases.
A Correlation of +1 indicates a strong positive relationship, one factor increases so does
the other.
“r”
Strong No Strong
Correlation Correlation Correlation
-1.0 0 +1.0
Decision Points
615
Linear Regression Review
F itte d L in e P lo t
Im p u r i ty = - 0 . 2 8 9 + 0 . 4 5 6 6 S tir r a te
2 0 .0 S 0.919316
R-S q 9 3.4 %
R - S q ( a d j) 9 2.7 %
1 7 .5
1 5 .0
Impurity
1 2 .5
The change in Y-value for
every one unit change in (X)
1 0 .0
Stirrate (Slope of the Line)
20 25 30 35 40 45
S t ir r a t e
616
Regression Analysis Review
Correlation only tells us the strength of a linear relationship, not the numerical
relationship.
The last step to proper analysis of Continuous Data is to determine the
regression equation.
The regression equation can mathematically predict Y for any given X.
Prediction Equations:
Y = a + bx (Linear or 1st order model)
Y = a + bx + cx2 (Quadratic or 2nd order model)
Y = a + bx + cx2 + dx3 (Cubic or 3rd order model)
Y = a (bx) (Exponential)
617
Simple vs. Multiple Regression Review
Simple Regression
– One X, One Y
– Analysis can be done with Data Analysis and
Scatter Plots
Multiple Regression
– Two or More X’s, One Y
– Analysis can be done with Data Analysis and
Scatter Plots
618
Regression Step Review
In doing Simple Linear Regression, you would find that residuals are not
normally distributed, or also find unusual observations.
These indicate departure from Simple Linear Regression (One X and One Y
with the Regression equation of the form, y=mx + c).
620
Curvilinear Regression
A sample data set for a contact center setting is as shown below. Note that the Call AHT is the eventual response variable (Y) and Call ACW is the
predictor variable we wish to investigate.
Steps for doing a Curvilinear Regression will remain the same with the only difference of one option.
Earlier, the equation was of the form, y = mx + c, but now it is y = m1x1 + m1x22 +c
One should use Curvilinear Regression only when Model Adequacy parameters are not met with Simple Linear Regression.
621
Types of Non-Linear Relationships
622
Multiple Linear Regression
What do you do when you have multiple input variables (Predictors) and one output variable (Response) to test?
Excel provides us a simple option of doing Multiple Linear Regression. All the steps of Simple Linear Regression hold except for drawing Scatter plots.
Drawing Scatter plots with multiple variables is a time consuming affair. You need to draw individual scatter plots for each variable. Yet, for illustration we
will adhere to the steps.
623
Multiple Linear Regression
Scatter charts for both the variables show a good fit line passing through all the points. Their Pearson’s coefficient of correlation is high as well.
Time for us to test out all other metrics and see how well does this model hold.
SUMMARY OUTPUT
Regression Statistics
Multiple R 1
R Square 1
Adjusted R Square 1
ANOVA
df SS MS F Significance F
Coefficients Standard Error t Stat P-value Lower 95% Upper 95% Lower 95.0% Upper 95.0%
624
Observation Predicted Y Residuals Percentile Y
Multiple Linear Regression
The Adjusted R-Square metric is close to 1, i.e. 100% of variability in residuals/ predicted values from actual values is explained by the model.
The p-values of the Intercept and one of the X variables are significant, thus demanding the need of inclusion of all the two in the final regression equation.
Call Wait time is taken off from the regression model because it showed a non-significant p-value.
Yet, when we did a simple scatter plot, we saw a Best Fit Line passing through the points in the Scatter.
In situations like these, help of process expertise is needed to help you decide if you need to consider both the variables for your final regression equation.
625
Transforming Process Data
Data that is asymmetric can often be transformed to make it more symmetric using a
numerical function which operates more strongly on large numbers than small ones;
such as logarithms and roots.
Transform Rules:
1. The transform must preserve the relative order of the data.
2. The transform must be a smooth and continuous function.
3. Most often useful when the ratio of largest to smallest value is greater than two. In most
cases, the transform will have little effect when this rule is violated.
4. All external reference points (spec limits, etc.) must use the same transform.
Transformation Power(p)
Cube 3
{ }
Square 2
xp
xtrans= No Change 1
log(x) Square Root 0.5
Logarithm 0
Reciprocal Root -0.5
Reciprocal -1
626
Effect of Transformation
20 15
Frequency
Frequency
15
10
10
5
5
0 0
10 20 30 40 50 60 70 80 90 100 0 10 20 30 40 50 60 70 80 90 100
Right Skew Sqrt
627
Transforming Data Using MINITABTM
628
Box Cox Transform
StDev
1.5
1.0
Limit
-1 0 1 2 3
Before Transform Lambda After Transform
Probability Plot of Pos skew Probability Plot of BoxCox
Normal Normal
99.9 99.9
Mean 1.050 Mean 0.9469
StDev 0.8495 StDev 0.3934
99 99
N 100 N 100
AD 2.883 AD 0.265
95 P-Value <0.005 95 P-Value 0.687
90 90
80 80
70 70
Percent
Percent
60 60
50 50
40 40
x 0.50 or x
30 30
20 20
10 10
5 5
1 1
0.1 0.1
-2 -1 0 1 2 3 4 5 6 0.0 0.5 1.0 1.5 2.0 2.5
Pos skew BoxCox
629
Transforming Without the Box Cox Routine
630
Transforming Without the Box Cox Routine
70
Percent
60
60
50
50
40
40
30
30
20
20
10
10
5
5
11
0.1
0.1
0.0
0.0 0.5
0.5 1.0
1.0 1.5
1.5 2.0
2.0 2.5
2.5
Square Root
Square Root
Our transform is the square root—the same as the Box Cox transform of
lambda = 0.5
631
Multiple Linear Regression
Regressions are run on historical process data. They are NOT a form of experimentation.
Multiple Linear Regression investigates multiple input variable’s effect on an output
simultaneously.
– If R2 is not as high as desired in the Simple Linear Regression.
– Process knowledge implies more than one input affects the output.
The assumptions for residuals with Simple Regressions are still necessary for Multiple Linear
Regressions.
An additional assumption for MLR is the independence of predictors (X’s).
– MINITABTM can test for multicollinearity (Correlation between the predictors or X’s).
Model error (residuals) is impacted by the addition of measurement error for all the input
variables.
632
Definitions of MLR Equation Elements
The definitions for the elements of the Multiple Linear Regression model are as follows:
633
MLR Step Review
634
Multiple Linear Regression Model Selection
635
Flight Regression Example
636
Flight Regression Example Matrix Plot
Matrix
Matrix Plot
Plot of
of Flight
Flight Speed,
Speed, Altitude,
Altitude, Turbine
Turbine Angl,
Angl, Fuel/Air
Fuel/Air rat,
rat, ...
...
600
600 750
750 900
900 32
32 36
36 40
40 99 12
12 15
15
Output Response
600
600
Flight
FlightSpeed
Speed 500
500
400
400
900
900
750
750 Altitude
Altitude
600
600
37.0
37.0
34.5
34.5
Turbine
TurbineAngle
Angle
32.0
32.0
40
40
36 Fuel/Air
36 Fuel/Airratio
ratio
32
32 19.5
19.5
18.0
18.0
ICR
ICR
16.5
16.5
15
15
12
12
Temp
Temp
99
400
400 500
500 600
600 32.0
32.0 34.5
34.5 37.0
37.0 16.5
16.5 18.0
18.0 19.5
19.5
Predictors
Since 2 or more predictors show Correlation, run MLR.
637
Flight Regression Example Best Subsets
F
T u
u e
r l
b /
i A
A n i
l e r
t
i A r
t n a T
u g t I e
Mallows d l i C m
Vars R-Sq R-Sq(adj) C-p S e e o R p
1 72.1 71.1 38.4 28.054 X
1 39.4 37.2 112.8 41.358 X
2 85.9 84.8 9.0 20.316 X X
2 82.0 80.6 17.9 22.958 X X
3 87.5 85.9 7.5 19.561 X X X
3 86.5 84.9 9.6 20.267 X X X
4 89.1 87.3 5.7 18.589 X X X X
4 88.1 86.1 8.2 19.481 X X X X
5 89.9 87.7 6.0 18.309 X X X X X
638
Flight Regression Example Model Selection
T
F
u
List of all the Predictors
u e (X’s)
r l
b /
i A
A n i
l e r
t
i A r
t n a T
u g t I e What model would you select?
Mallows d l i C m
Vars R-Sq R-Sq(adj) C-p S e e o R p
1 72.1 71.1 38.4 28.054 X Let’s consider the 5 predictor model:
1 39.4 37.2 112.8 41.358 X • Highest R-Sq(adj)
2 85.9 84.8 9.0 20.316 X X
2 82.0 80.6 17.9 22.958 X X • Lowest Mallows Cp
3 87.5 85.9 7.5 19.561 X X X
3 86.5 84.9 9.6 20.267 X X X
• Lowest S
4 89.1 87.3 5.7 18.589 X X X X • However there are many terms.
4 88.1 86.1 8.2 19.481 X X X X
5 89.9 87.7 6.0 18.309 X X X X X
639
Flight Regression Example Model Selection
Stat>Regression>Regression>Options
640
Flight Regression Example Model Selection
Predictor Coef SE Coef T P VIF The VIF for temp indicates it should
Constant 770.4 229.7 3.35 0.003 be removed from the model. Go back
Altitude 0.15318 0.06605 2.32 0.030 2.3
Turbine Angle 5.806 2.843 2.04 0.053 1.4
to the Best Subsets analysis and select
Fuel/Air ratio 8.696 3.327 2.61 0.016 3.2 the best model that does not include
ICR -52.269 6.157 -8.49 0.000 2.6 the predictor temp.
Temp 4.107 3.114 1.32 0.200 5.4
641
Flight Regression Example Model Selection
F
T u
u e
r l
b /
i A
A n i
l e r
t
i A r
t n a T
u g t I e
Mallows d l i C m Select a model with 4 terms
Vars R-Sq R-Sq(adj) C-p S e e o R p
1 72.1 71.1 38.4 28.054 X because Temp was removed as
1
2
39.4
85.9
37.2
84.8
112.8
9.0
41.358
20.316
X
X X
a predictor since it had
2 82.0 80.6 17.9 22.958 X X Correlation with the other
3 87.5 85.9 7.5 19.561 X X X
3 86.5 84.9 9.6 20.267 X X X variables.
4 89.1 87.3 5.7 18.589 X X X X Re-run the Regression.
4 88.1 86.1 8.2 19.481 X X X X
5 89.9 87.7 6.0 18.309 X X X X X
642
Flight Regression Example Model Selection
643
Flight Regression Example Model Selection
Regression Analysis: Flight Speed versus Turbine Angl, Fuel/Air rat, ICR
Re-run the
The P-value for Turbine Angle now
indicates it should be removed and Regression
re-run the Regression
because P > 0.05
644
Flight Regression Final Regression Model
645
Flight Regression Example Residual Analysis
646
Flight Regression Example Residual Analysis
Residual
Residual Plots
Plots for
for Flight
Flight Speed
Speed
Normal
Normal Probability
Probability Plot
Plot of
of the
the Residuals
Residuals Residuals
Residuals Versus
Versus the
the Fitted
Fitted Values
Values
99
99
Residual
StandardizedResidual
90 22
90
Percent
Percent
50
Standardized
50
00
10
10
11 -2
-2
-3.0
-3.0 -1.5
-1.5 0.0
0.0 1.5
1.5 3.0
3.0 450
450 500
500 550
550 600
600 650
650
Standardized Residual
Standardized Residual Fitted
FittedValue
Value
Histogram
Histogram of
of the
the Residuals
Residuals Residuals
Residuals Versus
Versus the
the Order
Order of
of the
the Data
Data
88
Residual
StandardizedResidual
22
66
Frequency
Frequency
Standardized
44
00
22
00 -2
-2
-2
-2 -1
-1 00 11 22 22 44 66 88 10
10 12
12 14
14 16
16 18
18 20
20 22
22 24
24 26
26 28
28
Standardized Residual
Standardized Residual Observation Order
Observation Order
647
Summary
648
Designing Experiments
DOE Methodology
Experimental Methods
649
Project Status Review
650
Six Sigma Strategy
p liers Cu u ts
p sto SIPOC
O
u
S me Inp
ut
VOC
Con rs
pu
Project Scope
trac Emplo
tors yees
st
P-Map, X-Y, FMEA
(X1) (X11) (X9)
(X2) (X3) (X4) (X8) Capability
(X6) (X7) (X5) (X10)
Box Plot, Scatter
(X3) (X4) (X1) (X11) Plots, Regression
(X5) (X8)
(X2)
Fractional Factorial
Full Factorial
(X5) (X3) Center Points
(X11)
(X4)
651
Reasons for Experiments
The Analyze Phase narrowed down the many inputs to a critical few, now it is necessary to
determine the proper settings for the vital few inputs because:
– The vital few potentially have interactions.
– The vital few will have preferred ranges to achieve optimal results.
– Confirm cause and effect relationships among factors identified in Analyze Phase (e.g.
regression)
Understanding the reason for an experiment can help in selecting the design and focusing
the efforts of an experiment.
Reasons for experimenting are:
– Problem Solving (Improving a process response)
– Optimizing (Highest yield or lowest customer complaints)
– Robustness (Constant response time)
– Screening (Further screening of the critical few to the vital few X’s)
Problem Solving
– Eliminate defective products or services.
– Reduce cycle time of handling transactional processes.
Optimizing
– Mathematical model is desired to move the process response.
– Opportunity to meet differing customer requirements (specifications or VOC).
Robust Design
– Provide consistent process or product performance.
– Desensitize the output response(s) to input variable changes including NOISE
variables.
– Design processes knowing which input variables are difficult to maintain.
Screening
– Past process data is limited or statistical conclusions
prevented good narrowing of critical factors in Analyze
Phase.
What are the differences between DOE modeling and physical models?
– A physical model is known by theory using concepts of physics, chemistry,
biology, etc...
– Physical models explain outside area of immediate project needs and include
more variables than typical DOE models.
– DOE describes only a small region of the experimental space.
654
Definition for Design of Experiments
DOE allows the experimenter to study the effect of many input variables that
may influence the product or process simultaneously, as well as possible
interaction effects (for example synergistic effects).
The goal of DOE is to find a design that will produce the information
required at a minimum cost.
One Factor at a Time (OFAT) is an experimental style but not a planned experiment
or DOE.
The graphic shows yield contours for a process that are unknown to the experimenter.
7 120 33 90
135 85
6
130 90
125
1 2 3 4 5 Optimum identified
95 with OFAT
120 7
– Full Factorials
• 2-5 input variables
Response
Surface
Full Factorial
Fractional Factorials
657
Nomenclature for Factorial Experiments
The general notation used to designate a full factorial design is given by:
658
Visualization of 2 Level Full Factorial
T P T*P
-1 -1 +1 Four experimental runs:
+1 -1 -1 • Temp = 300, Press = 500
-1 +1 -1 • Temp = 350, Press = 500
+1 +1 +1 • Temp = 300, Press = 600
Coded levels for factors • Temp = 350, Press = 600
659
Graphical DOE Analysis - The Cube Plot
This graph is used by the experimenter to visualize how the response data is
distributed across the experimental space.
3.35
3.35 1.50
1.50
11
Fulcrum
Fulcrum
2.10
2.10 0.90
0.90
-1
-1 -1
-1
-1
-1 11
Start
StartAngle
Angle
Catapult.mtw
661
Graphical DOE Analysis - The Main Effects Plot
This graph is used to see the relative effect of each factor on the output
response.
-1 1 -1 1
Fulcrum
5
2
-1 1
Avg Distance at Low Setting of Start Angle: 2.10 + 3.35 + 5.15 + 8.20 = 18.8/4 = 4.70
Main Effects Plot (data means) for Distance
-1 1 -1 1 -1 1
5.2
4.4
Dist
3.6
2.8
2.0
Start Angle Stop Angle Fulcrum
Avg. distance at High Setting of Start Angle: 0.90 + 1.50 + 2.40 + 4.55 = 9.40/4 = 2.34
Run # Start Angle Stop Angle Fulcrum Distance
1 -1 -1 -1 2.10
2 1 -1 -1 0.90
3 -1 1 -1 3.35
4 1 1 -1 1.50
5 -1 -1 1 5.15
6 1 -1 1 2.40
7 -1 1 1 8.20
8 1 1 1 4.55
663
Interaction Definition
Interactions occur when variables act together to impact the output of the process. Interactions
plots are constructed by plotting both variables together on the same graph. They take the form
of the graph below. Note that in this graph, the relationship between variable “A” and Y changes
as the level of variable “B” changes. When “B” is at its high (+) level, variable “A” has almost no
effect on Y. When “B” is at its low (-) level, A has a strong effect on Y. The feature of interactions
is non-parallelism between the two lines.
Higher
B-
Y
When B changes
from low to high, the
Output
output drops
When B changes dramatically.
from low to high, the
output drops very
B+
little. Lower
- A +
664
Degrees of Interaction Effect
Y B+
B+ Y B+ Y
B+
Low Low Low
- A + - A + - A +
Strong Interaction Moderate Reversal
High High
B- B-
Y Y
B+
B+ B+
Low Low
- A + - A +
665
Interaction Plot Creation
4.5
Mean
3.5
2.5
(4.55 + 2.40)/2 = 3.48
1.5
666
Graphical DOE Analysis - The Interaction Plots
22
on Y when Start
Stop
Stop
Angle is at its low AAngle
ngle
66
-1
-1
level. Stop
StopAAngle
ngle
11 44
22
667
Graphical DOE Analysis - The Interaction Plots
Interaction
InteractionPlot
Plot(data
(datameans)
means) for
for Distance
Distance
-1
-1 11
Start
Start
66
AAngle
ngle
-1
-1
44
Star 11
StarttAAngle
ngle
22
Stop
Stop
66
AAngle
ngle
-1
-1
44
Stop
StopAAngle
ngle
11
Choose this option
22
Fulcrum
for the additional
Fulcrum
plots.
66
-1
-1
11
44
Fulcr
Fulcrum
um
22
-1
-1 11 -1
-1 11
668
DOE Methodology
669
Generate Full Factorial Designs in MINITABTM
670
Create Three Factor Full Factorial Design
671
Create Three Factor Full Factorial Design
672
Create Three Factor Full Factorial Design
673
Three Factor Full Factorial Design
• When shown a Main Effects Plots and interactions, determine which effects and
interactions may be significant.
675
Experimental Methods
Designing Experiments
Methodology
676
DOE Methodology
677
Questions to Design Selection
678
Questions to Design Selection
Technical Considerations
What are the objectives/goals of the experiment:
1. What factors are important? (narrowed from Analyze Phase)
2. What is the operating range for each factor?
3. How can I minimize both the cost of DOE and the cost of running the
process?
4. How much change in the process do we require?
5. How close to optimal does the process currently run?
6. Are we tackling a centering or variation problem?
7. What is the impact to the process while running the DOE?
8. What are the costs of competing DOE designs?
9. What do you know about the process interactions?
679
DOE Methodology Step 1
• Relate how the experiment connects with the original project scope. Practically
speaking, what is this experiment supposed to accomplish?
1. Identify Root Cause
2. Measure Variation
3. Measure Output Response
• Have the measurement systems been verified for the Input Variables and Output
Response?
680
DOE Methodology Step 2
• Objective must include the critical characteristics and the desired outcome.
– If the experiment and project is tackling recurring issues, consider a different critical
characteristic.
• The characteristic may require a different physical phenomenon being measured
or with a differing measurement system.
• The measurement system precision and accuracy may influence the specific
output to be measured.
681
DOE Methodology Step 3
682
DOE Methodology Step 4
• Use the Analyze Phase and subject matter experts to select these factors.
• All factors must be independent of each other.
• Consider past results from previous experiments.
• Test the most likely candidates first.
• Factors not included in the designed experiment should be held constant and recorded.
• Noise or uncontrollable factors (typically environmental conditions) should be
monitored and the experimental design may be impacted (see Step 6).
The inputs selected by the team following the Six Sigma methodology are dwell time (sec), temperature
of solution (deg F) and concentration of solution (% solids). Noise factors of ambient temperature and
humidity were recorded and monitored.
683
DOE Methodology Step 5
684
DOE Methodology Step 5 (cont.)
“-” “+”
Factor Settings
685
DOE Methodology Step 5 (cont.)
686
DOE Methodology Step 5 (cont.)
Factor Settings
“-” “+”
The experiment is using coded levels:
Dwell time: +1 (20 sec); -1 (10 sec)
Temp of sol’n: +1 (80 deg F); -1 (100 deg F)
Conc. of sol’n: +1 (40%) ; -1 (20%)
687
DOE Methodology Step 6
688
DOE Methodology Step 6 (cont.)
Balanced and orthogonal designs are highly encouraged and the definition of
balanced and orthogonal is covered in a later module.
Center Points are used for investigating curvature and advanced designs. Center
Points are covered in a later module.
Blocking can be used to account for noise variables and is covered in a later
module.
689
DOE Methodology Step 6 (cont.)
690
DOE Methodology Step 6 (cont.)
For full
factorials, this
equals 2factors
Specified in
Step 2.
Typically 0.9
σ of process output
variable See first slide
of Step 6.
692
DOE Methodology Step 6 (cont.)
693
DOE Methodology Step 6 (cont.)
Additional considerations are required when determining what a sample size means.
For the experimental results to be representative of the process, sample across the
largest family of variation.
– It is also necessary to determine how to define a representative sample and
experimental unit.
• Characteristics of a representative sample are:
– Repeatable measurement and represents natural variation of the
process.
• An experimental unit is the basic unit to which an experimental run
can be applied and includes all the qualities of a representative sample.
694
DOE Methodology Step 6 (cont.)
Recall from the Analyze Phase the Multi-Vari tool described the three families of
variation. Consider these families of variation to determine how to sample with
replication for an experiment.
– Within Unit or Positional
• Within piece variation related to the geometry of the part.
• Variation across a single unit containing many individual parts such as a
wafer containing many computer processors.
• Location in a batch process such as plating.
– Between Unit or Cyclical
• Variation among consecutive pieces.
• Variation among groups of pieces.
• Variation among consecutive batches.
• Temporal or Over Time
• Shift-to-Shift
• Day-to-Day
• Week-to-Week
695
DOE Methodology Step 7
• Discuss the experimental scope, time and cost with the process owners prior to the
experiment.
• Some team members must be present during the entire experiment.
• After the experiment has started, are you getting output responses you expected?
– If not, quickly evaluate for Noise or other factors and consider stopping or
canceling the experiment.
• Use a log book to make notes of observations, other factor settings, etc.
• Communicate with the operators, technicians, staff about the experimental details and
why the experiment is being discussed before running the experiment.
– This communication can prevent “helping” by the operators, technicians, etc. that
might damage your experimental design.
• Alert the laboratory or quality technicians if your experiment will increase the number
of samples arriving during the experiment.
696
DOE Methodology Step 8
8. Analyze the Data from the Designed Experiment and draw Statistical
Conclusions
697
DOE Methodology Step 9
698
DOE Methodology Step 10
• After finding the Practical Results from Step 9, verify the results:
– Set the factors at the Practical Results found with Step 9 and see if the
process output responds as expected. This verification replicates the result
of the experiment.
– Do not forget your model has some error.
699
DOE Methodology Step 11
• If the objective of the experiment was accomplished and the Business Case is
satisfied, then proceed to the Control Plan which is covered in the Control Phase.
• Do not just run experiments and not implement the solutions.
• Further experiments may need to be designed to further change the output to
satisfy the Business Case.
– This possible need for another experiment is why we stated in earlier
modules that DOE’s can be an iterative process.
700
Summary
701
Full Factorial Experiments
702
Why Use Full Factorial Designs
703
Mathematical Output of Experiments
• The end result of a DOE is a mathematical function to describe the results of the
experiment.
• For the 2k Factorial designs this module discusses, Linear relationships are
covered.
• All models will have some error as shown by the ε in the below equation.
• The mathematical equation below is the Prediction from the experimental data.
Notice there is no error term in this form.
• is the predicted Output Response as a function of the Input Variables used in
Ŷ experiment.
the
704
Linear Mathematical Model
65
65
60 55
% Reacted % Reacted
1 1
55
45
-1
0
Cn -1
0
Cn
0 -1 0 -1
Ct 1 T 1
705
Quadratic Mathematical Model
Quadratic Models can be obtained with designs not described in this module.
Quadratic Models explain curvature, maximums, minimums and twisted maximums and
minimums when interactions are active.
– The following is the quadratic prediction model used in some response surface models
not covered in this training.
– The simpler 2k models do not include enough information to generate the Quadratic
Model.
Surface Plot of C6
21
16
C6
11
1.5
1.0
6 0.5
-1.5 -0.5
0.0
B
-1.0 -0.5 -1.0
0.0 0.5 -1.5
A 1.0 1.5
706
Nomenclature for Factorial Experiment
707
Treatment Combinations
Treatment combinations, or experimental runs, show how to set the levels for each
of the factors.
Minuses and plusses can be used to indicate low and high factor level settings,
Center Points are indicated with zeros.
Temperature
10 20 Treatment combination for
Pressure 50 1 2 run number 2 is:
100 3 4
Temperature at 20 deg and
Pressure at 50 psi.
708
Standard Order of 2 Level Designs
The design matrix for 2k factorials are shown in standard order (not randomized).
– The low level is indicated by a “-” and the high level by a “+”.
– This order is commonly referred to Yates standard order for Dr. Frank Yates.
709
Full Factorial Design with 4 Factors
710
Full Factorial Design
Factors
711
Balanced Design
An experiment is balanced when each factor has the same number of experimental
runs at both high and low levels.
A B
1 - -
2 + -
3 - +
4 + +
Xi 0 0
712
Orthogonal Design
AA BB CC AB
AB AC
AC BC
BC
11 -- -- ++ ++ -- --
22 ++ -- -- -- -- ++
33 -- ++ -- -- ++ --
44 ++ ++ ++ ++ ++ ++
Xi X y 00 00 00
713
Biomedical Production Example
In this example we will walk through the 11 Step DOE methodology. The
biomedical firm is attempting to increase the yield of a specific protein expression
for use in research by universities and pharmaceutical companies.
714
Biomedical Production Example
715
Biomedical Production Example
716
Biomedical Production Example
717
Biomedical Production Example
718
Biomedical Production Example
For ease of data entry for the results of the DOE, we have turned off “Randomize runs” by
removing the check mark in the window in the “Options…” tab.
719
Biomedical Production Example
In an empty column, C8, type in ‘Yield’ where we will place the experimental results.
Do NOT edit, copy, paste or alter anything in the first 7 columns or MINITAB TM will not
understand the worksheet.
720
Biomedical Production Example
721
Biomedical Production Example
722
Biomedical Production Example
723
Biomedical Production Example
Normal Probability Plot of the Standardized Effects The Normal Probability Plot assumes that
(response is Yield, Alpha = .05) insignificant effects are due to Noise and
99
Effect Ty pe therefore Normally Distributed. Any
Not Significant
95 Significant significant effects will be plotted off the
90 A F actor
A
N ame
Temp straight line and highlighted in red.
80 B C onc
AC C S upplier
70
Percent
60
50
40
30
20 Pareto Chart of the Standardized Effects
(response is Yield, Alpha = .05)
10
5 2.31
F actor N ame
A Temp
1
A B C onc
0 10 20 30 40 50 60 70 C S upplier
Standardized Effect AC
BC
effects graphically shows which
ABC
effects are significant based on the
selected alpha level. Any effect AB
724
Biomedical Production Example
In the Session Window under the Factorial Fit, any effect that has a P-value less than 0.05 (for
an alpha of 0.05) is considered significant.
Notice that all three methods of determining what effects belong in the final model fit agree.
725
Biomedical Production Example
726
Biomedical Production Example
727
Biomedical Production Example
728
Biomedical Production Example
Main Effects Plot (data means) for Yield Non-parallel lines in the
Temp Conc
Interaction Plot indicated
70
65 significance. The lines do not
60 have to cross each other to be
55
significant. Also, they can
Mean of Yield
50
25 45 5 15 cross slightly and still be
Supplier
insignificant.
70 Interaction Plot (data means) for Yield
65 5 15 A B
60 Temp
70 25
55 45
50 T emp 60
A B 50
C onc
70 5
15
729
Biomedical Production Example
730
Biomedical Production Example
Residual
Residual Plots
Plots for
for Yield
Yield
Normal
Normal Probability
Probability Plot
Plot of
of the
the Residuals
Residuals Residuals
Residuals Versus
Versus the
the Fitted
Fitted Values
Values
99
99 22
Residual
StandardizedResidual
90
90 11
Percent
Percent
Standardized
50
50 00
10 -1
-1
10
11 -2
-2
-2
-2 -1
-1 00 11 22 40
40 50
50 60
60 70
70 80
80
Standardized
StandardizedResidual
Residual Fitted
FittedValue
Value
Histogram
Histogram of
of the
the Residuals
Residuals Residuals
Residuals Versus
Versus the
the Order
Order of
of the
the Data
Data
44 22
Residual
StandardizedResidual
33 11
Frequency
Frequency
Standardized
22 00
11 -1
-1
00 -2
-2
-1.5
-1.5 -1.0
-1.0 -0.5
-0.5 0.0
0.0 0.5
0.5 1.0
1.0 1.5
1.5 11 22 33 44 55 66 77 88 99 10
10 11
11 12
12 13
13 14
14 15
15 16
16
Standardized Observation
ObservationOrder
StandardizedResidual
Residual Order
731
Biomedical Production Example
The Residuals versus variables are most important when deciding what level to set an
insignificant factor.
A typical guideline is a difference of a factor of 3 in the spread of the Residuals between the
low and high levels of an insignificant input variable.
– In this case concentration was not significant, but we still need to make a
decision on how to set it for the process. The low level for concentration has
a smaller spread of Residuals, but there is not a difference of 3:1. Other
considerations for setting the variable are cost and reducing cycle time.
Residuals Versus Temp Residuals Versus Conc
(response is Yield) (response is Yield)
2 2
Spread of residuals
1 1
Standardized Residual
Standardized Residual
0 0
-1 -1
-2 -2
25 30 35 40 45 5.0 7.5 10.0 12.5 15.0
Temp Conc
732
Biomedical Production Example
Practical Solution:
- Temp 45C
- Concentration 5%
- Supplier B
734
Biomedical Production Example
735
Center Points
A Center Point is an additional experimental run made at the physical center of the
design.
– Center Points do not change the model to quadratic.
– They allow a check for adequacy of Linear model.
The Center Point provides a check to see if it is valid to say that the output response
is Linear through the center of the design space.
If a straight line connecting high and low levels passes through the center of the
design, the model is adequate to predict inside the design space.
– “Curvature” is the statistic used to interpret the adequacy of the Linear Model.
– If curvature is significant
the P-value will be
less than 0.05. Output Response
Do NOT predict outside the
design space.
A Center Point is always a good insurance policy but is most effective when all the
input factors are Continuous.
A guideline is to run 2-4 Center Point runs distributed uniformly through the
experiment when all the input factors are continuous in a Full or Fractional
Factorial.
Y
Maximize Response
Does it matter that the
linear model is
inappropriate?
x
“-” “c” “+”
737
Panel Cleaning Example
In this example we will walk through the 11 step DOE methodology for a panel
cleaning machine using Center Points in the analysis. The manufacturing firm is
attempting to start up a new panel cleaning machine and would like to getting it
running quickly. They have experience with this type of machine, but they do not
have experience with this particular model of equipment.
738
Panel Cleaning Example
739
Panel Cleaning Example
Why are
Notice the Center
Center Points Points not
are uniformly random?
distributed
through the
design.
Panel Cleaning.mtw
740
Creating Designs with Center Points
MINITABTM will place the Center Points randomly in the worksheet. Through the
next few steps we will demonstrate how to move the Center Points so they are
uniformly distributed.
• Create a 3 factor design with 3 Center Points and 2 replicates, be sure to
randomize the design.
741
Creating Designs with Center Points
742
Creating Designs with Center Points
743
Creating Designs with Center Points
To complete the Center Point arrangement, sort the data on the RunOrder
column but DO NOT create a new worksheet.
Data>Sort
744
Panel Cleaning Example
745
Panel Cleaning Example
746
Panel Cleaning Example
90 C F actor N ame
A D w ell Time
80 B Temp
B C N a2S 2O 8
70
A
Percent
60
50
40
30
20
Pareto Chart of the Standardized Effects
10 BC (response is Width, Alpha = .05)
5
2.23
F actor N ame
1 A D w ell Time
-10 -5 0 5 10 15 20 C B Temp
Standardized Effect C N a2S 2O 8
B
Term BC
A
The significant effects are
AB
Na2S2O8, Temp, Dwell Time
AC
and the interaction of Temp
ABC
with Na2S2O8.
0 5 10 15 20
Standardized Effect
747
Panel Cleaning Example
748
Panel Cleaning Example
749
Panel Cleaning Example
Analysis of Variance for Width (coded units)3 DF for the 3 Main Effects, 1 DF for the
Source DF Seq SS Adj SS interaction
Adj MS effect
F in the model.
P
Main Effects 3 599.336 599.336 199.779 148.18 0.000
1 DF for curvature based on the
2-Way Interactions 1 95.111 95.111 95.111 70.55 0.000
difference between the average of the
Curvature 1 0.221 0.221 0.221 points
factorial 0.16and 0.692
the average of the
Residual Error 13 17.527 17.527 1.348 Points.
Center
Lack of Fit 3 6.669 6.669 2.223 2.05 0.171
Pure Error 10 10.858 10.858 13 DF for residual error broken into two
1.086
Total 18 712.195 components: Lack of Fit and Pure Error.
750
Panel Cleaning Example
Simply insert these values into the equation and do the math.
752
Panel Cleaning Example
36
34
Mean of Width
32
30
4 5 6 40 60 80
Na2S2O8
Interaction Plot (data means) for Width
38 40 60 80 1.8 2.1 2.4
36 Dwell
40
Time Point Type
34 4 C orner
Dwell T ime 5 C enter
32
32 6 C orner
30
24
1.8 2.1 2.4
Temp Point Type
40
40 C orner
60 C enter
T emp 80 C orner
32
753
Panel Cleaning Example
Cube
Cube Plot
Plot (data
(data means)
means) for
for Width
Width
CCenterpoint
enterpoint
36.875
36.875 43.350
43.350 Factorial
FactorialPoint
Point
33.245
33.245 38.395
38.395
80
80
35.020
35.020
Temp
Temp 36.010
36.010 41.000
41.000
2.4
2.4
Na2S2O8
Na2S2O8
23.025
23.025 25.895
25.895
40
40 1.8
1.8
44 66
Dwell
DwellTime
Time
754
Panel Cleaning Example
Residual
Residual Plots
Plots for
for Width
Width
Normal
NormalProbability
Probability Plot
Plot of
of the
the Residuals
Residuals Residuals
Residuals Versus
Versus the
the Fitted
Fitted Values
Values
99
99 22
Residual
StandardizedResidual
90
90 11
Percent
Percent
50 00
Standardized
50
-1
-1
10
10
11 -2
-2
-2
-2 -1
-1 00 11 22 20
20 25
25 30
30 35
35 40
40
Standardized Residual
Standardized Residual Fitted Value
Fitted Value
Histogram
Histogramof
of the
the Residuals
Residuals Residuals
Residuals Versus
Versus the
the Order
Order of
of the
the Data
Data
22
6.0
Residual
6.0
StandardizedResidual
4.5 11
4.5
Frequency
Frequency
00
Standardized
3.0
3.0
1.5 -1
-1
1.5
0.0 -2
-2
0.0
-2
-2 -1-1 00 11 22 22 44 66 88 10
10 1212 1414 16
16 18
18
Standardized
StandardizedResidual
Residual Observation
ObservationOrder
Order
755
Panel Cleaning Example
1 1
Standardized Residual
Standardized Residual
0 0
-1 -1
-2 -2
4.0 4.5 5.0 5.5 6.0 40 50 60 70 80
Dwell Time Temp
1
Standardized Residual
-1
-2
1.8 1.9 2.0 2.1 2.2 2.3 2.4
Na2S2O8
756
Panel Cleaning Example
757
Panel Cleaning Example
The Response Optimizer has a little trick; if you include Center Points in the model
it will treat the low, center and high values as discrete points.
As you can see the Center Points fit the Linear Model.
Composite
Desirability
0.99885
Width
Targ: 40.0
y = 40.0057
d = 0.99885
758
Panel Cleaning Example
Composite
Desirability
1.0000
Width
Targ: 40.0
y = 40.0
d = 1.0000
759
Panel Cleaning Example
Width
Targ: 40.0
y = 40.0
d = 1.0000
Predicted output
760
Panel Cleaning Example
761
Panel Cleaning Example
762
Panel Cleaning Example
Na2S2O8
Na2S2O8
2.10
Dwell Time at 2.10 Dwell Time at
low setting middle setting
1.95 1.95
1.80 1.80
40 50 60 70 80 40 50 60 70 80
Temp Temp
shown in 45
Hold Values
2.25
white are the
Dwell Time 6
solution set
Na2S2O8
Sodium 1.80
Persulfate. 40 50 60
Temp
70 80
763
Panel Cleaning Example
764
Summary
Explain how Fit & Diagnose & Center Points factors into an experiment
765
Fractional Factorial Experiments
Designing Experiments
Experimental Methods
Designs
Full Factorial Experiments
Creation
Fractional Factorial Experiments
Generators
766
Why Use Fractional Factorial Designs
767
Why Use Fractional Factorial Designs
768
Nomenclature for Fractional Factorials
5-1
V
The example clarifies how to use the nomenclature.
• How many factors in the experiment?
• How many runs if no repeats or replicates?
• What fractional design is this (1/8, 1/4 or 1/2)?
769
Half-Fractional Experiment Creation
Recall the 2x2x2 full 3-factor, 2-level Factorial Design. Suppose we needed to investigate a fourth factor
but we could NOT increase the number of runs because of time or cost. Select the highest order interaction
to represent the levels of the fourth factor. The ABC interaction will determine the levels for factor D.
When we replace the ABC interaction with factor D, we say the ABC 3-way interaction was aliased or
confounded with D. This experiment maintains balance and orthogonality.
– The first experimental run in the first row indicates the experiment is executed with factor D at the low level
while running all the 3 other factors at the low level.
Factor D
A B C AxB AxC BxC AxBxC
-1 -1 -1 1 1 1 -1
1 -1 -1 -1 -1 1 1
-1 1 -1 -1 1 -1 1
1 1 -1 1 -1 -1 -1
-1 -1 1 1 -1 -1 1
1 -1 1 -1 1 -1 -1
-1 1 1 -1 -1 1 -1
1 1 1 1 1 1 1
Why is the design, shown as orange rows, called a “half” fraction? This is the design just created on
the previous slide. This is a half fraction since a full 2x2x2x2 factorial would take 16 runs. With the
half fraction we can estimate the effects of 4 factors in 8 runs. What is the cost? We lose the ability to
study the higher order interaction independently!
Could we create a quarter fraction experiment out of the above matrix and
still study four factors at once?
Why or why not?
771
Graphical Representation of Half-Fraction
- A + - A +
Top line of previous slide
-
- C
+
B
-
+ C
+
- D +
Remember that D is confounded with the ABC interaction in this half-fractional
design.
772
Design Generators
773
Design Generators
A B C AB AC BC D AD BD CD
-1 -1 -1 1 1 1
1 -1 -1 -1 -1 1
-1 1 -1 -1 1 -1
1 1 -1 1 -1 -1
-1 -1 1 1 -1 -1
1 -1 1 -1 1 -1
-1 1 1 -1 -1 1
1 1 1 1 1 1
774
Design Generators
A B C AB AC BC D AD BD CD
-1 -1 -1 1 1 1 -1 1 1 1
1 -1 -1 -1 -1 1 1 1 -1 -1
-1 1 -1 -1 1 -1 1 -1 1 -1
1 1 -1 1 -1 -1 -1 -1 -1 1
-1 -1 1 1 -1 -1 1 -1 -1 1
1 -1 1 -1 1 -1 -1 -1 1 -1
-1 1 1 -1 -1 1 -1 1 -1 -1
1 1 1 1 1 1 1 1 1 1
775
MINITABTM Session Window
A + BCD
B + ACD
C + ABD
D + ABC
AB + CD
AC + BD
AD + BC
776
So What is “Confounding”?
777
Confounded Effects With Fractionals
778
Experimental Resolution
Resolution IV
Next hold up four fingers
The Confounding is main effects with three
way interactions or…
Main Effects Three Way Interactions
k-p
The visual aid is shown through Resolution V.
R
Resolution V
Hold up Five Fingers, One on one hand and
Four on the other. This illustrates the
Confounding of main effects with four way
Main Effects Four Way Interactions interactions or …
780
MINITABTM Fractional Factorial Design Creation
Fortunately, MINITABTM creates the designs for us to prevent having to create a Fractional
Factorial by hand. This output, found in the MINITABTM session window after creating a
Fractional Factorial design, should be understood because it also informs us of the Resolution
of the design.
A + BCD
B + ACD
C + ABD
D + ABC
AB + CD
AC + BD
AD + BC
781
(5 -1)
2V Fractional Design Resolution V
Example of a very useful Fractional Design often used for screening designs.
Run A B C D E
1 -1 -1 -1 -1 1
2 1 -1 -1 -1 -1
3 -1 1 -1 -1 -1
E
4 1 1 -1 -1 1
5 -1 -1 1 -1 -1
6 1 -1 1 -1 1
7 -1 1 1 -1 1
B
8 1 1 1 -1 -1
C
A 9 -1 -1 -1 1 -1
D 10 1 -1 -1 1 1
11 -1 1 -1 1 1
Pros Cons 12 1 1 -1 1 -1
13 -1 -1 1 1 1
5 factors (Main Effects) 16 trials to get 5 Main Effects 14 1 -1 1 1 -1
10 2-way interactions 2nd order interactions are 15 -1 1 1 1 -1
Main Effects only Confounded with Confounded with 3rd order
rare 4-way interactions 16 1 1 1 1 1
782
MINITABTM’s Display of Available Designs
Note: Since we discourage Design Resolution III or IV, MINITABTM has shaded these as
RED and YELLOW for cautionary. GREEN is acceptable because Main Effects are not
Confounded with lower level interactions.
783
DOE Methodology
784
Fractional Factorial Example
785
Fractional Factorial Example (cont.)
786
Fractional Factorial Example (cont.)
787
Fractional Factorial Example (cont.)
788
Fractional Factorial Example (cont.)
Pareto
Pareto Chart
Chart of
of the
the Effects
Effects
(response
(response isis Y,
Y, Alpha
Alpha =
= .10)
.10)
0.26
0.26
EE FFactor
actor NName
ame
AA AA
AAC
C BB BB
HH CC CC
BB DD DD
AAFF EE EE
FF FF
AAEE GG GG
AAD
D HH HH
Term
Term
AA
AAG
G
CC
AAH
H
AABB
GG
FF
DD
00 22 44 66 88 10
10 12
12 14
14
Effect
Effect
Lenth's
Lenth's PSE
PSE =
= 0.129375
0.129375
789
Fractional Factorial Example (cont.)
A choice must be made in reducing our model or reducing the number of terms in the model. We
have chosen to look at the Confounding table generated by MINITAB TM.
The AC 2 factor interaction is Confounded with other 2-way interactions but we will
assume for now using the Confounding table from MINITAB TM that the 2-way AC
interaction is actually the EH 2 factor interaction because both factors E and H are
significant.
The second highest effect for a 2 factor interaction AF. We will look at the
Confounding table and assume it is the BE 2-way interaction since the B and E factors
are significant.
The 2-way interaction AE also is significant with the alpha above 0.1. We cannot find
another 2-way interaction that might be significant using just the B, E, and H factors.
If the AE interaction is kept in the model, then to maintain “hierarchical order”
factors A and E must be kept in the model.
We will now lower reduce the model and see if we can further reduce the model.
790
Fractional Factorial Example (cont.)
The Reduced Model is shown here and we want 95% confidence to include terms
Notice the AE 2-way interaction has the smallest effect of the statistically significant terms and
factor A kept in the model to maintain the “hierarchical order” also has a small term and is
statistically insignificant. We choose to reduce the model and remove those terms. R-sq should
not be severely impacted. If it was impacted severely, we would reconsider this choice.
791
Fractional Factorial Example (cont.)
792
Fractional Factorial Example (cont.)
Residual
Residual Plots
Plots for
for YY
Normal
NormalProbability
Probability Plot
Plot Residuals
Residuals Versus
Versus the
the Fitted
Fitted Values
Values
99
99 NN 16
16 0.4
0.4
AD
AD 0.532
0.532
90
90 P-Value
P-Value 0.146
0.146 0.2
0.2
Residual
Percent
Residual
Percent
50
50
0.0
0.0
10
10 -0.2
-0.2
11
-0.50
-0.50 -0.25
-0.25 0.00
0.00 0.25
0.25 0.50
0.50 00 10
10 20
20 30
30
Residual
Residual Fitted
FittedValue
Value
Histogram
Histogram of
of the
the Residuals
Residuals Residuals
Residuals Versus
Versus the
the Order
Order of
of the
the Data
Data
44 0.4
0.4
33 0.2
Frequency
0.2
Frequency
Residual
Residual
22
0.0
0.0
11
-0.2
-0.2
00
-0.3
-0.3 -0.2
-0.2 -0.1
-0.1 0.0
0.0 0.1
0.1 0.2
0.2 0.3
0.3 0.4
0.4 11 22 33 44 55 66 77 88 99 10
10 11
11 12
12 13
13 14
14 15
15 16
16
Residual Observation
ObservationOrder
Order
Residual
793
Fractional Factorial Example (cont.)
No, no unusual
observations here…
794
Fractional Factorial Example (cont.)
795
Fractional Factorial Example (cont.)
It can be difficult to optimize the solutions and get the Practical Solution desired.
Using Response Optimizer within MINITAB TM helps us find the Practical Solution of setting the
factors left in the model all at the high level or +1
796
Fractional Factorial Example (cont.)
797
Fractional Factorial Example (cont.)
We win, we win…!!
798
Fractional Factorial Example (cont.)
799
Fractional Factorial Exercise
800
Summary
801
Control Phase
Advanced Capability
Welcome to Control
Capability as Monitoring Technique
803
Beginnings of Control Phase
You’ve already narrowed to the “vital few” with the Define, Measure, Analyze and
Improve Phases.
Just because you are able to achieve results with your project or through a DOE
does not mean you have Process Capability.
This module in the Control Phase gives you tools and ideas to tackle Special Causes
that may be hampering your Process Capability even if you found your “vital few”
to get an improved average.
804
Capability and Monitoring
• If the project was important enough to warrant the time and attention of
you and your team, it is important enough to ensure that performance
levels are maintained
• Monitoring the improved process is a key element of the Control Plan
• Reporting Capability and Stability should be used together as primary
components of the monitoring plan
• In the Measure Phase, Capability was used to establish baseline
performance by assessing what had occurred in the relevant past
• In the Control Phase, Capability becomes a predictive (inferential) tool to
predict the expected process performance, usually based on a sample.
805
Capability Studies
Capability Studies
• Are intended to be periodic estimations of a process’s ability to meet its
requirements
• Can be conducted on both Discrete and Continuous Data
• Are most meaningful when conducted on stable processes
• Can be reported as Sigma Level which is optimal (short term) performance
• These concepts should be remembered from using the Six Sigma toolset applied
so far:
– Customer or business specification limits
• Business specification limits cannot be wider than the specification limits
of a final product
– Nature of long term vs. short term data
– Mean and Standard Deviation of the process (for Continuous Data)
– The behavior and shape of the distribution of Continuous Data
– Procedure for determining Sigma level
– Relevance of data
806
Components of Variation
As in the Measure Phase, understanding whether you are dealing with long
term or short term data is an important first step.
If the process is stable, short term data provides a quick estimate of true
process potential since special causes are minimal.
Lot 1 Lot 5
Lot 3
Lot 2
Lot 4
Short-term studies
Long-term study
807
Capability for the Control Phase
Assess Stability
Verify Specifications
N Are Data
Continuous?
N Determine the
Defects or Are Data
Shape of the
D efe ctive s
Defectives? Normal?
Distribution
D efe cts
Stat> Quality Tools> Stat> Quality Tools> Stat> Quality Tools> Stat> Quality Tools>
808
Stability Check
• To assess the stability of the performance, you need to use a tool called Control Charts.
• Control Charts vary on the type of data for the metric chosen.
• For defectives data, you should plot p chart or np chart. For defects per unit data, you should plot u chart or c chart.
In this example, we wish to track if the late coming of reports out of the total class is stable or not.
Perform the standard calculations for a p-chart using the formulas provided by the facilitator
Day Subgroup size Late Reports
1 89 14
2 101 18
3 108 13
4 80 12
5 107 13
6 77 15
7 76 18
8 101 18
9 72 16
10 105 12
11 102 14
12 82 15
13 92 16
14 76 15
15 105 20
16 99 11
17 76 20
18 105 11
19 88 19
20 78 16
21 108 17
22 89 14
23 80 15
24 105 19
25 102 20
26 105 18 Is the process stable in behavior?
27 80 31
28 89 16
29 88 12
30 97 17
809
Control Phase
Discrete Capability: Poisson Output
Our target was to be less than 0.05 DPU. Based on the Excel data sheet used to make the p –chart and calculations shown below, the DPU
is less than 0.05.
USL 0.3
Then to calculate Sigma levels, you can either use the Sigma tables or use
Defective per unit 0.026316 the formula
=-NORMSINV(DPU) + 1.5
RTY 97.40%
810
Normal Capability Sixpack
Use the data sheet provided to you in the LSSBB Data set in the worksheet
mentioned Control Chart Data.
Step 1 – Calculate Sample Means and Sample Ranges for the data sample chosen.
Step 3 --- Use the Control Chart formulas for both the Xbar and R Charts, as below
For R Chart
811
Normal Capability Sixpack
Use the data sheet provided to you in the LSSBB Data set in the worksheet
mentioned Control Chart Data.
812
Lean Controls
813
Lean Controls
You’ve begun the process of sustaining your project after finding the “vital few”
X’s for your project.
This module gives more tools from the Lean toolbox to stabilize your process.
Belts, after some practice, often consider this module’s set of tools a way to
improve some processes that are totally “out of control” or of significantly poor
Process Capability before applying the Six Sigma methodology.
814
The Vision of Lean Supporting Your Project
Kanban
The Continuous Goal… We cannot sustain Kanban
Sustaining Results Kaizen without Kaizen.
Standardized Work
We cannot sustain Kaizen (Six
Sigma) without Standardized Work.
Waste is often the root of any Six Sigma project. The 7 basic elements of waste (muda
in Japanese) include:
– Muda of Correction
– Muda of Overproduction
– Muda of Processing
– Muda of Conveyance
– Muda of Inventory
– Muda of Motion
– Muda of Waiting
Get that garbage outta here!
– The reduction of MUDA can reduce your outliers and help with defect
prevention. Outliers because of differing waste among procedures, machines, etc.
816
The Goal
Don’t forget the goal -- Sustaining your Project which eliminates MUDA!
With this in mind, we will introduce and review some of the Lean tools used to sustain
your project success.
817
5 S - Workplace Organization
818
5 S Translation - Workplace Organization
819
SORTING - Decide what is needed.
Definition:
– To sort out necessary and
unnecessary items.
– To store often used items at the
work area, infrequently used
items away from the work area
and dispose of items that are not
needed.
Why: Things
Thingstotoremember
remember
•• Start
Startininone
onearea,
area,then
thensort
sortthrough
through
– Removes waste. everything.
everything.
– Safer work area. •• Discuss
Discussremoval
removalofofitems
itemswith
withall
all
persons involved.
persons involved.
– Gains space. •• Use
Useappropriate
appropriatedecontamination,
decontamination,
– Easier to visualize the process. environmental
environmentaland andsafety
safetyprocedures.
procedures.
•• Items
Itemsthat
thatcannot
cannotbe
beremoved
removed
immediately
immediately should betagged
should be taggedforforlater
later
removal.
removal.
•• IfIfnecessary,
necessary,use
usemovers
moversandandriggers.
riggers.
820
A Method for Sorting
Item
Keep &
Monitor
Keep &
Store
821
A Method for Sorting
A
B
C
Distance
822
STRAIGHTENING – Arranging Necessary Items
Definition:
– To arrange all necessary items.
– To have a designated place
for everything.
– A place for everything and everything in its
place.
– Easily visible and accessible.
Why:
– Visually shows what is required or is out of Things
place. Thingstotoremember
remember
•• Things
Thingsusedusedtogether
togethershould
shouldbe
be
– More efficient to find items and documents kept together.
(silhouettes/labels). kept together.
•• Use
Uselabels,
labels,tape,
tape,floor
floor
– Saves time by not having to
markings,
markings,signs,
signs,and
and
search for items.
shadow
shadowoutlines.
outlines.
– Shorter travel distances.
•• Sharable
Sharableitems
itemsshould
shouldbe bekept
kept
atataacentral location (eliminated
central location (eliminated
excess).
excess).
823
SHINING – Cleaning the Workplace
Definition:
– Clean everything and find
ways to keep it clean.
– Make cleaning a part of your
everyday work.
Why:
– A clean workplace indicates
a quality product and
process. Things
Thingstotoremember
remember
– Dust and dirt cause product •• “Everything
“Everythingininits itsplace”
place”frees
freesupuptime
timefor
for
contamination and potential cleaning.
cleaning.
health hazards. •• Use
Useananoffice
officeororfacility
facilitylayout
layoutasasaavisual
visual
– A clean workplace helps aid to identify individual responsibilities
aid to identify individual responsibilities
identify abnormal for
forcleaning.
cleaning.This
Thiseliminates
eliminates“no “noman’s
man’s
conditions. land.”
land.”
•• Cleaning
Cleaningthe thework
workareaareaisislike
likebathing.
bathing.ItIt
relieves
relievesstress
stressand
andstrain,
strain,removes
removessweat
sweat
and
and dirt, and prepares the body for thenext
dirt, and prepares the body for the next
day.
day.
824
STANDARDIZING – Creating Consistency
Definition:
– To maintain the workplace at a
level that uncovers problems
and makes them obvious.
– To continuously improve your
office or facility by continuous
assessment and action.
Why:
– To sustain Sorting, Storage and
Things
Thingstotoremember
remember
Shining activities every day. • • We
Wemust
mustkeep
keepthe
thework
workplace
placeneat
neatenough
enough
for visual identifiers to be effective in
for visual identifiers to be effective in
uncovering
uncoveringhidden
hiddenproblems.
problems.
• • Develop
Developaasystem
systemthat
thatenables
enableseveryone
everyoneinin
the
theworkplace
workplacetotosee
seeproblems
problemswhen
whenthey
they
occur.
occur.
825
SUSTAINING – Maintaining the 5S
Definition:
– To maintain our discipline,
we need to practice and repeat
until it becomes a way of life.
Why:
Things
ThingstotoRemember
Remember
– To build 5S into our everyday •• Develop
Developschedules
schedulesand
andcheck
check
process. lists.
lists.
•• Good
Goodhabits
habitsare
arehard
hard
totoestablish.
establish.
•• Commitment
Commitmentand anddiscipline
discipline
toward
toward housekeepingare
housekeeping areessential
essential
first steps toward being world
first steps toward being world
class.
class.
826
The Visual Factory
A Visual Factory enables a process to manage its processes with clear indications of opportunities.
Your team should ask the following questions if looking for a project:
– Can we readily identify Downtime Issues?
– Can we readily identify Scrap Issues?
– Can we readily identify Changeover Problems?
– Can we readily identify Line Balancing Opportunities?
– Can we readily identify Excessive Inventory Levels?
– Can we readily identify Extraneous Tools & Supplies?
Exercise:
– Can you come up with any opportunities for “VISUAL” aids in your project?
– What visual aids exist to manage your process?
827
What is Standardized Work?
We cannot sustain
Standardized Work without
5S and the Visual Factory.
5S - Workplace Organization
828
Prerequisites for Standardized Work
Standardized work does not happen without the visual factory which can be further
described with:
Visual alert of variation in the process (visual factory). Operators, material handlers,
office staff all need visual signals to keep “standard work” a standard
Identified and labeled in-process stock (5S). As inventory levels of in-process stock
decrease, a visual signal should be sent to the material handlers to replenish this stock
829
What is Kaizen?
Measurable Process. Without standardized work, we really wouldn’t have a consistent process to
measure. Cycle times would vary, assembly methods would vary, batches of materials would be
mixed, etc…
Analysis Tools. There are improvement projects in each organization which cannot be solved by
an operator. This is why we teach the analysis tools in the breakthrough strategy of Six Sigma.
Operator Support. The organization needs to understand that its future lies in the success of the
value-adding employees. Our roles as Belts are to convince operators that we are here for them--
they will then be there for us.
831
What is Kanban?
Kanbans are the best control method of inventory which impacts some of the 7
elements of MUDA shown earlier.
Kanban provides production, conveyance, and delivery information. Kanban
In it’s purest form the system will not allow any goods to be
moved within the facility without an appropriate Kanban (or
signal) attached to the goods.
– The Japanese word for a communication signal
Kaizen
or card--typically a signal to begin work
– Kanban is the technique Standardized Work
used to “pull” products and
material through and into
the lean manufacturing system.
– The actual “Kanban” can be a
physical signal such as an empty Visual Factory
container or a small card.
5S - Workplace Organization
832
Two Types of Kanban
Supplier
833
Prerequisites for a Successful Kanban System
• Number of parts per Kanban (card) MUST be standard and SHOULD be kept to as few
as possible parts per card.
• Near zero defects should be sent to the assembly process (Result of earlier belt projects).
834
Warnings Regarding Kanban
835
The Lean Tools and Sustained Project Success
The Lean tools help sustain project success. The main lessons you should consider are:
1. The TEAM should 5S the project area and begin integrating visual factory indicators.
– Indications of the need for 5S are:
– Outliers in your project metric
– Loss of initial gains from project findings
3. Kaizen’s and Kanban’s cannot be attempted without organized workplaces and organized
work instructions.
– Remember the need for 5S and Standardized Work Instructions to support our
projects.
4. Project Scope dictates how far up the Lean tools ladder you need to implement measures
to sustain any project success from your DMAIC efforts.
836
Class Exercise
In the boundaries for your project scope, give some examples of Lean tools in operation.
List other Lean tools you are most interested in applying to sustain your project results.
837
Summary
• Understand how tools must document the defect prevention created in the Control
Phase
838
Defect Controls
Advanced Capability
839
Purpose of Defect Prevention in Control Phase
840
Level for Project Sustaining in Control
5-6Six Sigma product and/or process design eliminates an error condition OR an BEST
automated system monitors the process and automatically adjust critical X’s to
correct settings without human intervention to sustain process improvements
3-5: Mistake proofing prevents a product/service from passing onto the next step
3-4: SPC on X’s with the special causes are identified and acted upon by fully
trained operators and staff who adhere to the rules
2-4SPC on Y’s
WORST
841
6 Product/Process Design
Designing products and processes such that the output Y meets or exceeds the target
capability.
24
Specification on Y
22
Distribution 21
of Y
19
Relationship Y
17 = F(x)
10 11 12 13 14 15 16 17 18 19 20
Distribution of X
When designing the part or process, specifications on X are set such that the target capability
on Y is achieved.
Both the target and tolerance of the X must be addressed in the spec limits.
842
6s Product/Process Design
Upper
Prediction
24
Interval
Specification on Y
22
Distribution 21 Relationship Y
of Y
= F(x)
19
17
10 11 12 13 14 15 16 17 18 19 20 Lower
Prediction
Distribution of X Interval
843
Product/Process Design Example
80
70
60
50
Output
40
30
What are the 20
spec limits for 10
Regression
the output? 0
95% PI
0 5 10
Input
What is the tolerance range for the input?
If you want 6 performance, you will remember to tighten the output’s specification
to select the tolerance range of the input.
844
Product/Process Design Example
Regression Plot
Y = 2.32891 - 0.282622X
R-Sq = 96.1 %
10
Regression
Input2 90
80
70
60
Output
50
40
30
20
Regression
0 5 10
Input
Using top output spec determines high or low tolerance for input
depending on slope of Regression.
845
Poor Regression Impacting Tolerancing
Regression Plot
Y = -4.7E-01 R-Sq =
+ 0.811312X 90.4 % Poor Correlation does not allow
for tighter tolerancing.
20
Outp1
10
20
Outp2
10
Regression
0
95% PI
0 10 20 30
Inp1
846
5 – 6 Full Automation
Full Automation: Systems that monitor the process and automatically adjust critical
X’s to correct settings.
847
Full Automation Example
As part of the Control Phase, the team has implemented an automatic check and
replenish system on the washer.
Full Automation
Process Interruption: Mechanism installed that shuts down the process and prevents
further operation until a required action is preformed:
• Ground fault circuit breakers
• Child proof caps on medications
• Software routines to prevent undesirable commands
• Safety interlocks on equipment such as light curtains, dual palm buttons, ram
blocks
• Transfer system guides or fixtures that prevent over or undersized parts from
proceeding
• Temperature conveyor interlocks on ovens
• Missing component detection that stops the process when triggered
849
4 – 5 Process Interruption
Example:
• A Black Belt is working on launching a new electric drive unit on a transfer system
– One common failure mode of the system is a bearing failure on the main motor
shaft
– It was determined that a high press fit at bearing installation was causing these
failures
– The root cause of the problem turned out to be undersized bearings from the
supplier
• Until the supplier could be brought into control or replaced, the team implemented a
press load monitor at the bearing press with a indicator
– If the monitor detects a press load higher than the set point, it shuts down the
press and will not allow the unit to be removed from press until an interlock key
is turned and the ram reset in the manual mode
– Only the line lead person and the supervisor have keys to the interlock
– The non-conforming part is automatically marked with red dye
Process Interruption
850
3 – 5 Mistake Proofing
851
Traditional Quality vs. Mistake Proofing
Traditional Inspection
Result
Sort
Worker or Don’t Do Defective At Other
Machine Error Anything Step
Source Inspection
“KEEP ERRORS FROM
TURNING INTO DEFECTS”
852
Styles of Mistake Proofing
There are 2 states of a defect which are addressed with mistake proofing.
SHUTDOWN SHUTDOWN
(Stop Operation) (Stop Operation)
853
Mistake Proofing Devices Design
854
Types of Mistake Proof Devices
Contact Method
– Physical or energy contact
with product
1 Guide Pins of
Different Sizes
• Limit switches
• Photo-electric beams
Fixed Value Method 2 Error Detection
and Alarms
– Number of parts to be
attached/assembled etc.
constant
are
3 Limit Switches
Motion-step Method
– Checks for correct sequencing
– Checks for correct timing 5 Checklists
• Photo-electric switches
and timers
855
Mistake Proofing Examples
856
Advantages of Mistake Proofing as A Control Method
857
Defect Prevention Culture and Good Control Plans
858
Class Exercise
Break into your groups and discuss mistake proofing systems currently at your facilities
Identify one automation example and one process interruption example per group
859
Class Exercise about Defect Prevention
860
Summary
• Understand what tools must document the Defect Prevention created in the Control Phase
861
Statistical Process Control
Advanced Capability
Lean Controls
Elements and Purpose
Defect Controls
Methodology
Statistical Process Control (SPC)
Special Cause Tests
Examples
862
SPC Overview: Collecting Data
Population:
– An entire group of objects that have been made or will be made
containing a characteristic of interest
Sample:
– A sample is a subset of the population of interest
– The group of objects actually measured in a statistical study
– Samples are used to estimate the true population parameters
Population
Sample
Sample
Sample
863
SPC Overview: I-MR Chart
• An I-MR chart combines a Control Chart of the average moving range with the Individual’s Chart.
• You can use individuals charts to track the process level and to detect the presence of Special Causes when the
sample size is 1.
• Seeing both charts together allows you to track both the process level and process variation at the same time,
providing greater sensitivity that can help detect the presence of Special Causes.
I-MR Chart
U C L=226.12
225.0
Individual Value
222.5
_
220.0 X=219.89
217.5
215.0
LC L= 213.67
1 13 25 37 49 61 73 85 97 109
O bse r v a tion
8
U C L=7.649
6
Moving Range
4
__
M R=2.341
2
0 LC L= 0
1 13 25 37 49 61 73 85 97 109
O bse r v a tion
864
SPC Overview: Xbar-R Chart
If each of your observations consists of a subgroup of data, rather than just individual measurements, an
Xbar-R chart providers greater sensitivity. Failure to form rational subgroups correctly will make your
Xbar-R charts dangerously wrong.
Xbar-R Chart
U C L=225.76
225
Sample Mean
222 _
_
X=221.13
219
LC L=216.50
216
1 3 5 7 9 11 13 15 17 19 21 23
Sample
U C L=16.97
16
Sample Range
12
_
8 R=8.03
0 LC L=0
1 3 5 7 9 11 13 15 17 19 21 23
Sample
865
SPC Overview: U Chart
U Chart of Defects
0.14 1
1
UCL=0.1241
0.12
Sample Count Per Unit
0.10
0.08
0.06 _
U=0.0546
0.04
0.02
0.00 LCL=0
1 3 5 7 9 11 13 15 17 19
Sample
866
SPC Overview: P Chart
P Chart of Errors
0.30
UCL=0.2802
0.25
Proportion
_
0.20 P=0.2038
0.15
LCL=0.1274
1 3 5 7 9 11 13 15 17 19
Sample
867
SPC Overview: Control Methods/Effectiveness
Type 1 Corrective Action = Countermeasure: improvement made to the process which will eliminate
the error condition from occurring. The defect will never be created. This is also referred to as a long-
term corrective action in the form of mistake proofing or design changes.
Type 2 Corrective Action = Flag: improvement made to the process which will detect when the error
condition has occurred. This flag will shut down the equipment so that the defect will not move
forward.
SPC on X’s or Y’s with fully trained operators and staff who respect the rules. Once a chart signals a
problem everyone understands the rules of SPC and agrees to shut down for Special Cause
identification. (Cpk > certain level).
SPC on X’s or Y’s with fully trained operators. The operators have been trained and understand the
rules of SPC, but management will not empower them to stop for investigation.
S.O.P. is implemented to attempt to detect the defects. This action is not sustainable short-term or long-
term.
868
Purpose of Statistical Process Control
SPC is used to detect Special Cause variation telling us the process is “out of control”
but does NOT tell us why.
SPC gives a glimpse of ongoing process capability AND is a visual management tool.
869
Elements of Control Charts
UCL=55.24
Special Cause 50
Variation Detected 40
Individual Value
30
_
X=29.06 Process Center
(usually the Mean)
20
Control Limits
10
LCL=2.87
0
1 4 7 10 13 16 19 22 25 28
Observation
870
Understanding the Power of SPC
Control Charts indicate when a process is “out of control” or exhibiting Special Cause variation but NOT why!
SPC Charts allow workers and supervision to maintain improved process performance from Six Sigma projects.
Control Limits describe the process variability and are unrelated to customer specifications. (Voice of the Process
instead of Voice of the Customer)
– An undesirable situation is having Control Limits wider than customer specification limits. This will exist
for poorly performing processes with a Cp less than 1.0
Many SPC Charts exist and selection must be appropriate for effectiveness.
871
The Control Chart Cookbook
872
Focus of Six Sigma and the Use of SPC
Y=F(x)
To get results, should we focus our behavior on the Y or X?
Y X1 . . . XN
Dependent Independent
Output Input
Effect Cause
Symptom Problem
Monitor Control
Special Cause
Variation Process Run Chart of data
is “Out of Control” points
Upper Control
Limit
Common Cause
+/- 3 sigma
Variation Process
is “In Control”
Lower Control
Limit
Mean
Special Cause
Variation Process
is “Out of Control”
Process Sequence/Time Scale
874
Control and Out of Control
Outlier
3
2
1
99.7%
95%
68%
-1
-2
-3
Outlier
875
Size of Subgroups
Lot 1 Lot 5
Lot 3
Lot 2
Lot 4
Short-term studies
Long-term study
876
The Impact of Variation
-UCL
-LCL
If you base your limits on all three sources of variation, what will sound the alarm?
877
Frequency of Sampling
Sampling Frequency is a balance between cost of sampling and testing versus cost of not
detecting shifts in Mean or variation.
Process knowledge is an input to frequency of samples after the subgroup size has been
decided.
– If a process shifts but cannot be detected because of too infrequent sampling, the
customer suffers
– If choice is given of large subgroup samples infrequently or smaller subgroups
more frequently, most choose to get information more frequently.
– In some processes, with automated sampling and testing frequent sampling is
easy.
A rule of thumb also states “sample a process at least 10X more frequent than the
frequency of ‘out of control’ conditions”.
878
Frequency of Sampling
Sampling too little will not allow for sufficient detection of shifts in the process because
of Special Causes.
I Chart of Sample_3
Output 7.5
UCL=7.385
Individual Value
6.5
_
6.5 6.0
X=6.1
6
5.5 5.5
5 5.0
Sample every half hour LCL=4.815
1 7 13 19 25 31 37
1 2 3 4 5 6 7 8 9 10 11 12 13
Observation
6.2
7
Individual Value
Individual Value
6.0
_ _
X=6.129 X=5.85
6 5.8
5.6
5
Sample every hour 5.4
Sample 4x per shift
5.2
LCL=5.141
4 LCL=4.090
5.0
1 2 3 4 5 6 7 1 2 3 4
Observation Observation
879
SPC Selection Process
Choose Appropriate
Control Chart
type of
subgroup
attribute
size
data
DEFECTS DEFECTIVES
Cumulative Exponentially
Sum Weighted Moving
Average
880
Understanding Variable Control Chart Selection
Average & Range or S Production is higher volume; allows process Mean and variability to be viewed and
(Xbar and R or assessed together; more sampling than with Individuals Chart (I) and Moving Range
Xbar and S) Charts (MR) but when subgroups are desired. Outliers can cause issues with Range (R)
charts so Standard Deviation charts (S) used instead if concerned.
Most common
Individual and Production is low volume or cycle time to build product is long or homogeneous sample
Moving Range represents entire product (batch etc.); sampling and testing is costly so subgroups are
not desired. Control limits are wider than Xbar Charts. Used for SPC on most inputs.
Pre-Control Small shift needs to be detected, often because of autocorrelation of the output results.
Used only for individuals or averages of Outputs. Infrequently used because of
Exponentially calculation complexity.
Weighted
Moving Average Same reasons as EWMA (Exponentially Weighted Moving Range) except the past data
is as important as present data.
Cumulative Sum
Less Common
881
Understanding Attribute Control Chart Selection
882
Detection of Assignable Causes or Patterns
Control Charts indicate Special Causes being either assignable causes or patterns.
The following rules are applicable for both variable and Attribute Data to detect Special Causes.
These four rules are the only applicable tests for Range (R), Moving Range (MR) or Standard Deviation (S) charts.
– One point more than 3 Standard Deviations from the Center Line.
– 6 points in a row all either increasing or all decreasing.
– 14 points in a row alternating up and down.
– 9 points in a row on the same side of the center line.
These remaining four rules are only for variable data to detect Special Causes.
– 2 out of 3 points greater than 2 Standard Deviations from the Center Line on the same side.
– 4 out of 5 points greater than 1 Standard Deviation from the Center Line on the same side.
– 15 points in a row all within one Standard Deviation of either side of the Center Line.
– 8 points in a row all greater than one Standard Deviation of either side of the Center Line.
883
Recommended Special Cause Detection Rules
• If implementing SPC manually without software initially, the most visually obvious violations are more
easily detected. SPC on manually filled charts are common place for initial use of defect prevention
techniques.
• These 3 rules are visually the most easily detected by personnel.
– One point more than 3 Standard Deviations from the Center Line.
– 6 points in a row all either increasing or all decreasing.
– 15 points in a row all within one Standard Deviation of either side of the Center Line.
• Dr. Shewhart that worked with the Western Electric Co. was credited with the following 4 rules referred
to as Western Electric Rules.
– One point more than 3 Standard Deviations from the Center Line.
– 8 points in a row on the same side of the Center Line.
– 2 out of 3 points greater than 2 Standard Deviations from the Center Line on the same side.
– 4 out of 5 points greater than 1 Standard Deviation from the Center Line on the same side.
• You might notice the Western Electric rules vary slightly. The importance is to be consistent in your
organization and decide what rules you will use to detect Special Causes.
• VERY few organizations use all 8 rules for detecting Special Causes.
884
Special Cause Rule Default in MINITABTM
If a Belt is using MINITABTM, you must be aware of what default settings for the
rules. You can alter your program defaults with:
Tools>Options>Control Charts and Quality Tools> Tests
Many experts have commented on the appropriate tests and numbers to be used.
Decide then be consistent when implementing.
885
Special Cause Test Examples
This is the MOST common Special Cause test used in SPC charts.
A
B
C
C
B
A
886
Special Cause Test Examples
A
B
C
C
B 2
887
Special Cause Test Examples
B
C
C
B
A
888
Special Cause Test Examples
A
B
C
C 4
B
A
889
Special Cause Test Examples
890
Special Cause Test Examples
891
Special Cause Test Examples
B
A
892
Special Cause Test Examples
893
SPC Center Line and Control Limit Calculations
Calculate the parameters of the Individual and MR Control Charts with the following:
R
k
x i i
UC L x X E 2M R U C L MR D 4M R
X i 1 MR i
k k LC L x X E 2M R LC L MR D 3M R
Where:
Xbar: Average of the individuals, becomes the Center Line on the Individuals Chart
Xi: Individual data points
k: Number of individual data points
Ri : Moving range between individuals, generally calculated using the difference between each
successive pair of readings
MRbar: The average moving range, the Center Line on the Range Chart
UCLX: Upper Control Limit on Individuals Chart
LCLX: Lower Control Limit on Individuals Chart
UCLMR: Upper Control Limit on moving range
LCLMR : Lower Control Limit on moving range (does not apply for sample sizes below 7)
E2, D3, D4: Constants that vary according to the sample size used in obtaining the moving range
894
SPC Center Line and Control Limit Calculations
Calculate the parameters of the Xbar and R Control Charts with the following:
R
k
x i i UC L x X A 2R UC L R D 4R
X i 1
R i
LC L x X A 2R LC L R D 3R
k k
Where:
Xi: Average of the subgroup averages, it becomes the Center Line of the Control Chart
Xi: Average of each subgroup
k: Number of subgroups
Ri : Range of each subgroup (Maximum observation – Minimum observation)
Rbar: The average range of the subgroups, the Center Line on the Range Chart
UCLX: Upper Control Limit on Average Chart
LCLX: Lower Control Limit on Average Chart
UCLR: Upper Control Limit on Range Chart
LCLR : Lower Control Limit Range Chart
A2, D3, D4: Constants that vary according to the subgroup sample size
895
SPC Center Line and Control Limit Calculations
Calculate the parameters of the Xbar and S Control Charts with the following:
x i s i UC L x X A 3S UC L S B 4S
X i 1
S i 1
k k LC L x X A 3S LC L S B 3S
Where:
Xi: Average of the subgroup averages, it becomes the Center Line of the Control Chart
Xi: Average of each subgroup
k: Number of subgroups
si : Standard Deviation of each subgroup
Sbar: The average S. D. of the subgroups, the Center Line on the S chart
UCLX: Upper Control Limit on Average Chart
LCLX: Lower Control Limit on Average Chart
UCLS: Upper Control Limit on S Chart
LCLS : Lower Control Limit S Chart
A3, B3, B4: Constants that vary according to the subgroup sample size
896
SPC Center Line and Control Limit Calculations
LC L c c 3 c
Where:
900
SPC Center Line and Control Limit Calculations
Calculate the parameters of the EWMA Control Charts with the following:
902
Pre-Control Charts
Pre-Control Charts use limits relative to the specification limits. This is the first and ONLY
chart you will see specification limits plotted for Statistical Process Control. This is the most
basic type of chart and unsophisticated use of process control.
Qualifying Process
• To qualify a process, five consecutive parts must fall within the green zone
• The process should be qualified after tool changes, adjustments, new operators,
material changes, etc
904
Responding to Out of Control Indications
• The power of SPC is not to find out what the Center Line and Control Limits are.
• The power is to react to the Out of Control (OOC) indications with your Out of Control Action Plans
(OCAP) for the process involved. These actions are your corrective actions to correct the output or
input to achieve proper conditions.
40
1
UCL=39.76
VIOLATION:
Special Cause is indicated
30
Individual Value
20 _
X=18.38
10 OCAP
If response time is too high, get
0
LCL=-3.01
additional person on phone bank
1 4 7 10 13 16 19 22 25 28 31
Observation
905
Attribute SPC Example
Practical Problem: A project has been launched to get rework reduced to less than
25% of paychecks. Rework includes contacting a manager about overtime hours to
be paid. The project made some progress but decides they need to implement SPC to
sustain the gains and track % defective. Please analyze the file “paycheck2.mtw” and
determine the Control Limits and Center Line.
Step 3 and 5 of the methodology is the primary focus for this example.
– Select the appropriate Control Chart and Special Cause tests to employ
– Calculate the Center Line and Control Limits
– Looking at the data set, we see 20 weeks of data.
– The sample size is constant at 250.
– The amount of defective in the sample is in column C3.
Paycheck2.mtw
906
Attribute SPC Example (cont.)
907
Attribute SPC Example (cont.)
We will confirm what rules for Special Causes are included in our Control
Chart analysis.
908
Attribute SPC Example (cont.)
Remember to click on the “Options…” and “Tests” tab to clarify the rules for
detecting Special Causes.
…. Chart Options>Tests
We will confirm what rules for Special Causes are included in our Control
Chart analysis. The top 3 were selected.
909
Attribute SPC Example (cont.)
UCL=0.2802
0.25
Proportion
_
0.20 P=0.2038
0.15
LCL=0.1274
1 3 5 7 9 11 13 15 17 19
Sample
Now we must see if the next few weeks are showing Special Cause from the
results. The sample size remained at 250 and the defective checks were 61, 64,
77.
910
Attribute SPC Example (cont.)
Remember, we have calculated the Control Limits from the first 20 weeks. We must
now put in 3 new weeks and NOT have MINITABTM calculate new Control Limits
which will be done automatically if we do not follow this technique. We are executing
Steps 6-8
– Step 6: Plot process X or Y on the newly created Control Chart
– Step 7: Check for Out-Of-Control (OOC) conditions after each point
– Step 8: Interpret findings, investigate Special Cause variation, & make
improvements following the Out of Control Action Plan (OCAP)
911
Attribute SPC Example (cont.)
…… Chart Options>Parameters
P Chart of Empl_w_Errors
1
0.30
UCL=0.2802
0.25
The new updated SPC chart is
Proportion
shown with one Special Cause. 0.20
_
P=0.2038
0.15
LCL=0.1274
1 3 5 7 9 11 13 15 17 19 21 23
Sample
912
Attribute SPC example (cont.)
Because of the Special Cause, the process must refer to the OCAP or Out of Control Action Plan that states what
Root Causes need to be investigated and what actions are taken to get the process back in Control.
P Chart of Empl_w_Errors
1
0.30
UCL=0.2802
0.25
Proportion
_
0.20 P=0.2038
0.15
LCL=0.1274
1 3 5 7 9 11 13 15 17 19 21 23
Sample
After the corrective actions were taken, wait until the next sample is taken to see if the process has changed to not
show Special Cause actions.
– If still out of control, refer to the OCAP and take further action to improve the process. DO NOT make
any more changes if the process shows back in control after the next reading.
• Even if the next reading seems higher than the Center Line! Don’t cause more variability.
If process changes are documented after this project was closed, the Control Limits should be recalculated as in step
9 of the SPC methodology.
913
Variable SPC Example
Practical Problem: A job shop drills holes for its largest customer as a
final step to deliver a highly engineered fastener. This shop uses five drill
presses and gathers data every hour with one sample from each press
representing a subgroup. The data is gathered in columns C3-C7.
Step 3 and 5 of the methodology is the primary focus for this example.
– Select the appropriate Control Chart and Special Cause tests to employ
– Calculate the Center Line and Control Limits
Holediameter.mtw
914
Variable SPC Example (cont.)
915
Variable SPC Example (cont.)
Specifications were never discussed. Let us calculate the Control Limits and
Center Line for this example.
We will confirm what rules for Special Causes are included in our Control
Chart analysis.
916
Variable SPC Example (cont.)
Remember to click on the “Options…” and “Tests” tab to clarify the rules for
detecting Special Causes.
……..Xbar-R Chart Options>Tests
We will confirm what rules for Special Causes are included in our Control
Chart analysis. The top 2 of 3 were selected.
917
Variable SPC Example (cont.)
Also confirm the Rbar method is used for estimating Standard Deviation.
Stat>Control Charts>Variable Charts for Subgroups>Xbar-R>Xbar-R Chart Options>Estimate
918
Variable SPC Example (cont.)
No Special Causes were detected in the Xbar Chart. The average hole diameter was
26.33. The UCL was 33.1 and 19.6 for the LCL.
Xbar-R
Xbar-R Chart
Chartof
of Part1,
Part1, ...,
..., Part5
Part5
35
35
UUCCL=33.07
L=33.07
ean
30
SampleMMean
30
_
__
_
X=26.33
Sample
X=26.33
25
25
20 LC
20 LCL=19.59
L=19.59
11 66 11
11 16
16 21
21 26
26 31
31 36
36 41
41 46
46
Sample
Sample
1
1
24 UUCCL=24.72
L=24.72
24
Range
18
SampleRange
18
__
12
12 R=11.69
R=11.69
Sample
66
00 LC
LCL=0
L=0
11 66 11
11 16
16 21
21 26
26 31
31 36
36 41
41 46
46
Sample
Sample
Now we will use the Control Chart to monitor the next 2 hours and see if we are still
in control.
919
Variable SPC Example (cont.)
Remember, we have calculated the Control Limits from the first 20 weeks. We must
now put in 2 more hours and NOT have MINITABTM calculate new Control Limits
which will be done automatically if we do not follow this step. We are executing Steps
6-8
– Step 6: Plot process X or Y on the newly created Control Chart
– Step 7: Check for Out-Of-Control (OOC) conditions after each point
– Step 8: Interpret findings, investigate special cause variation, & make
improvements following the Out of Control Action Plan (OCAP)
920
Variable SPC Example (cont.)
Sample Mean
30
Sample Range
18
_
12 R=11.69
0 LC L=0
1 6 11 16 21 26 31 36 41 46 51
Sample
921
Variable SPC Example (cont.)
Because of no Special Causes, the process does not refer to the OCAP or Out of Control Action Plan
and NO actions are taken.
Xbar-R
Xbar-RChart
Chartof
ofPart1,
Part1,...,
...,Part5
Part5
35
35
U C L=33.07
U C L=33.07
ean
30
SampleMMean
30
_
__
_
X=26.33
Sample
X=26.33
25
25
20 LC L=19.59
20 LC L=19.59
1 6 11 16 21 26 31 36 41 46 51
1 6 11 16 21 26 31 36 41 46 51
Sample
Sample
1
1
24 U C L=24.72
24 U C L=24.72
Range
18
SampleRange
18
_
12 _
R=11.69
12 R=11.69
Sample
6
6
0 LC L=0
0 LC L=0
1 6 11 16 21 26 31 36 41 46 51
1 6 11 16 21 26 31 36 41 46 51
Sample
Sample
If process changes are documented after this project was closed, the Control Limits should be
recalculated as in Step 9 of the SPC methodology.
922
Recalculation of SPC Chart Limits
923
SPC Chart Option in MINITABTM for Levels
Remembering many of the tests are based on the 1st and 2nd Standard Deviations from
the Center Line, some Belts prefer to have some additional lines displayed. This is
possible with:
The extra lines can be helpful if users are using MINITABTM for the SPC.
924
Summary
925