Professional Documents
Culture Documents
Submitted To Submitted by
Letter of Transmittal
Acknowledgement
Operation Management is the study of that part of the organization which is responsible for producing
goods and /or service.
Our Report on management of quality and quality control. In this regard our group is working on it for
some weeks, I personally the leader of our group .and would like to thanks to our honorable course teacher
Md. Asiqur Rahman (Assistant professor). He makes us interested to know something about the quality of
product and services and its silent features.
At last, I want to give thanks to Allah for his grace on us in preparing the Repot Successfully.
Page |4
Table of contents
07 Six Sigma 15
12 P charts 23-25
13 C charts 25-27
Management of quality
And
Quality control
##Although they may vary somewhat from product to product, or between a product and
a service, generally speaking, the dimensions of quality include:
Performance-main characteristics of the product or service.
Aesthetics-appearance, feel, smell, taste.
Special features-extra characteristics.
Conformance-how well a product or service corresponds to design specifications, and
to the customer's expectations.
Safety-risk of injury or harm.
Reliability-consistency of performance.
Durability-the useful life of the product or service.
Perceived quality-indirect evaluation of quality (e.g., reputation).
Service after sale-handling of complaints or checking on customer satisfaction.
It is important for management to recognize the different ways that the quality of a firm's products or
services can affect the organization and to take these into account in developing and maintaining a quality
assurance program. Some of the major ways that quality affects an organization are:
1. Loss of business
2. Liability
3. Productivity
4. Costs
Loss of business:
Poor designs or defective products or services can result in loss of business. Failure to devote adequate
attention to quality can damage a profit-oriented organization's image and lead to a decreased share of the
market, or it can lead to increased criticism and/or controls for a government agency or nonprofit
organization.
Liability:
Organizations must pay special attention to their potential liability due to damages or injuries resulting from
either faulty design or poor workmanship. This applies to both products and services. Thus, a poorly
designed steering arm on a car might cause the driver to lose control of the car, but so could improper
assembly of the steering arm. However, the net result is the same.
Productivity:
Productivity and quality are often closely related. Poor quality can adversely affect productivity during the
manufacturing process if parts are defective and have to be reworked or if an assembler has to try a number
of parts before finding one that fits properly.
Costs:
Poor service can mean having to redo the service and reduce service productivity. Conversely, improving
and maintaining good quality can have a positive effect on productivity.
Page |9
Any serious attempt to deal with quality issues must take into account the costs associated with quality.
Those costs can be classified into four categories:
1. Appraisal costs.
2. Prevention costs.
3. Internal failure costs.
4. External failure costs.
Appraisal costs:
Cost of activities designed to ensure quality or uncover defects, appraisal costs relate to inspection, testing,
and other activities intended to uncover defective product or services.
Prevention costs:
Prevention cost relate to attempt to prevent defects from occurring, they include costs such as planning and
administration systems, working with vendor, training, quality control procedure.
Internal failure costs:
Internal failure costs are those discovered during the production process, such as Scarp, rework,
investigation cost, lost production time.
External failure costs:
External failure costs are those discovered after delivery to the customer. External failure for a variety of
reason including defective material from vendors, incorrect machine settings, fault equipment.
Quality Certification
ISO 9000
The purpose of the International Organization for Standardization (ISO) is to promote worldwide standards
that will improve operating efficiency, improve productivity, and reduce costs. The ISO is composed of the
national standards bodies of 91 countries. The U.S. representative body is the American National Standards
Institute (ANSI). The work of the ISO is conducted by some 180 technical committees. ISO 9000 is the
work of the Quality Management and Quality Assurance Committee.
ISO 9000 is a set of international standards on quality management and quality assurance critical to
international business.
ISO 9000 certification and registration process are particularly helpful for companies that do not currently
have a quality management system; it provides the guideline for establishing the system and making
effective in it.
ISO 9000 standards include the following categories
System requirement.
Management requirement.
Resource requirement.
Realization requirement.
Remedial requirement.
Eight quality management principles formed the basis of the latest version, ISO 9000 2000:
The quality management principles are:
1. A systems approach to management.
2. Continual improvement.
3. Factual approach to decision making.
4. Mutually beneficial supplier relationships.
5. Customer focus.
6. Leadership.
7. People involvement.
8. Process approach.
P a g e | 11
ISO 14000
ISO 14000 is a set of international standards for assessing company’s environmental performance,
Proponents hope that the ISO 14000 will receive the same high interest that the ISO 9000 standards have
in the international business community, and will result in businesses giving more attention to
environmental responsibilities.
The standards for ISO 14000 certification bear upon Three major area:
Management systems: Systems development and integration of environmental responsibilities into
business planning.
Operations: Consumption of natural resources and energy.
Environmental systems: Measuring, assessing, and managing emissions, effluents, and other waste
streams.
Proponents hope that the ISO 14000 will receive the same high interest that the ISO 9000 standards have
in the international business community, and will result in businesses giving more attention to
environmental responsibilities.
The preceding description provides a good idea of what TQM is all about, but it doesn’t tell the whole story.
A number of other elements of TQM are important:
1. Continuous improvement. The philosophy that seeks to improve all factors related to the process
of converting inputs into outputs on an ongoing basis is called continuous improvement. It covers
equipment, methods, materials, and people. Under continuous improvement, the old adage “If it
ain’t broke, don’t fix it” gets transformed into “Just because it isn’t broke doesn’t mean it can’t be
improved.”
2. Competitive benchmarking. This involves identifying other organizations that are the best at
something and studying how they do it to learn how to improve your operation. The company need
not be in the same line of business.
3. Employee empowerment. Giving workers the responsibility for improvements and the authority
to make changes to accomplish them provides strong motivation for employees. This puts decision
making into the hands of those who are closest to the job and have considerable insight into
problems and solutions.
4. Team approach. The use of teams for problem solving and to achieve consensus takes advantage
of group synergy, gets people involved, and promotes a spirit of cooperation and shared values
among employees.
5. Decisions based on facts rather than opinions. Management gathers and analyzes data as a basis
for decision making.
6. Knowledge of tools. Employees and managers are trained in the use of quality tools.
7. Supplier quality. Suppliers must be included in quality assurance and quality improvement efforts
so that their processes are capable of delivering quality parts and materials in a timely manner.
8. Champion. A TQM champion’s job is to promote the value and importance of TQM principles
throughout the company.
9. Quality at the source. Quality at the source refers to the philosophy of making each worker
responsible for the quality of his or her work. The idea is to “Do it right the first time.” Workers
are expected to provide goods or services that meet specifications and to find and correct mistakes
that occur. In effect, each worker becomes a quality inspector for his or her work. When the work
is passed on to the next operation in the process (the internal customer) or, if that step is the last in
the process, to the ultimate customer, the worker is “certifying” that it meets quality standards.
10. Suppliers are partners in the process, and long-term relationships are encouraged. This gives
suppliers a vital stake in providing quality goods and services. Suppliers, too, are expected to
provide quality at the source, thereby reducing or eliminating the need to inspect deliveries from
suppliers.
It would be incorrect to think of TQM as merely a collection of techniques. Rather, TQM reflects a whole
new attitude toward quality. It is about the culture of an organization. To truly reap the benefits of TQM,
the organization must change its culture.
P a g e | 13
Companies have had varying success in implementing TQM. Some have been quite successful, but others
have struggled. Part of the difficulty may be with the process by which it is implemented rather than with
the principles of TQM. Among the factors cited in the literature are the following:
1. Lack of a companywide definition of quality: Efforts aren’t coordinated; people are working at
cross-purposes, addressing different issues, and using different measures of success.
2. Lack of a strategic plan for change: Without such a plan the chance of success is lessened and
the need to address strategic implications of change is ignored.
3. Lack of a customer focus: Without a customer focus, there is a risk of customer dissatisfaction.
4. Poor intraorganizational communication: The left hand doesn’t know what the right hand is
doing; frustration, waste, and confusion ensue.
5. Lack of employee empowerment: Not empowering employees gives the impression of not
trusting employees to fix problems, adds red tape, and delays solutions.
6. View of quality as a “quick fix”: Quality needs to be a long-term, continuing effort.
7. Emphasis on short-term financial results: “Duct-tape” solutions often treat symptoms; spend a
little now—a lot more later.
8. Inordinate presence of internal politics and “turf” issues: These can sap the energy of an
organization and derail the best of ideas.
9. Lack of strong motivation: Managers need to make sure employees are motivated.
10. Lack of time to devote to quality initiatives: Don’t add more work without adding additional
resources.
11. Lack of leadership: Managers need to be leaders.
This list of potential problems can serve as a guideline for organizations contemplating implementing
TQM or as a checklist for those having trouble implementing it.
Criticisms of TQM
TQM programs are touted as a way for companies to improve their competitiveness, which is a very
worthwhile objective. Nonetheless, TQM programs are not without criticism. The following are some of
the major criticisms:
1. Overzealous advocates may pursue TQM programs blindly, focusing attention on quality even
though other priorities may be more important (e.g., responding quickly to a competitor’s
advances).
2. Programs may not be linked to the strategies of the organization in a meaningful way.
3. Quality-related decisions may not be tied to market performance. For instance, customer
satisfaction may be emphasized to the extent that its cost far exceeds any direct or indirect benefit
of doing so.
4. Failure to carefully plan a program before embarking on it can lead to false starts, employee
confusion, and meaningless results.
5. Organizations sometimes pursue continuous improvement (i.e., incremental improvement) when
dramatic improvement is needed.
6. Quality efforts may not be tied to results.
Note that there is nothing inherently wrong with TQM; the problem is how some individuals or
organizations misuse it. Let’s turn our attention to problem solving and process improvement.
Six Sigma
Six sigma A business process for improving quality, reducing costs, and increasing customer satisfaction.
Conceptually, the term is much broader, referring to a program designed to reduce the occurrence of defects
to achieve lower costs and improved customer satisfaction. It is based on the application of certain tools
and techniques to selected projects to achieve strategic business results. In the business world, six-sigma
programs have become a key way to improve quality, save time, cut costs, and improve customer
satisfaction. Six-sigma programs can be employed in design, production, service, inventory management,
and delivery. It is important for six-sigma projects to be aligned with organization strategy.
There are management and technical components of six-sigma programs. The management component
involves providing strong leadership,
Defining performance metrics,
Selecting projects likely to achieve business results,
Selecting and training appropriate people.
The technical component involves improving process performance,
Reducing variation,
Utilizing statistical methods,
Designing a structured improvement strategy,
which involves definition, measurement, analysis, improvement, and control.
Lean/Six Sigma
Lean/six sigma is a balanced approach to process improvement that integrates principles from lean
operation and statistical tools for variation reduction from six sigma to achieve speed and quality. Lean/six
sigma combines the power of the workers, who are close to the process, and the structured approach of six-
sigma methodology. Managers facilitate the creative problem-solving process. The rationale of this
approach is that lean principles alone cannot achieve statistical process control, and six sigmas alone cannot
achieve improved process speed and flow. The approach is equally applicable to products and services. In
fact, some of the earliest applications were in service support functions of General Electric and Caterpillar
Finance.
P a g e | 16
The plan-do-study-act (PDSA) cycle, also referred to as either the Shewhart cycle or the Deming wheel, is
the conceptual basis for problem-solving activities.
P a g e | 17
Plan.: Begin by studying the current process. Document that process. Then collect data on the process or
problem. Next, analyze the data and develop a plan for improvement. Specify measures for evaluating the
plan.
Do: Implement the plan, on a small scale if possible. Document any changes made during this phase. Collect
data systematically for evaluation.
Study: Evaluate the data collection during the do phase. Check how closely the results match the original
goals of the plan phase.
Act: If the results are successful, standardize the new method and communicate the new method to all
people associated with the process. Implement training for the new method. If the results are unsuccessful,
revise the plan and repeat the process or cease this project.
Quality Control
Process Variability:
All processes generate output that exhibits some degree of variability. The issue is whether the output
variations are within an acceptable range. This issue is addressed by answering two basic questions about
the process variations:
1. Are the variations random? If nonrandom variations are present, the process is considered to be
unstable. Corrective action will need to be taken to improve the process by eliminating the causes
of nonrandomness and achieve a stable process.
2. Given a stable process, is the inherent variability of process output within a range that conforms to
performance criteria? This involves assessment of a process’s capability to meet standards. If a
process is not capable, this situation will need to be addressed.
The natural or inherent process variations in process output are referred to as chance or random variations.
Such variations are due to the combined influences of countless minor factors, each one so unimportant that
even if it could be eliminated, the impact on process variations would be negligible. In Deming’s terms,
this is referred to as common variability. The amount of inherent variability differs from process to process.
For instance, older machines generally exhibit a higher degree of natural variability than newer machines,
partly because of worn parts and partly because new machines may incorporate design improvements that
lessen the variability in their output. A second kind of variability in process output is called assignable
variation, or nonrandom variation. In Deming’s terms, this is referred to as special variation. Unlike natural
variation, the main sources of assignable variation can usually be identified (assigned to a specific cause)
and eliminated. Tool wear, equipment that needs adjustment, defective materials, human fac-tors
(carelessness, fatigue, noise and other distractions, failure to follow correct procedures, and so on) and
problems with measuring devices are typical sources of assignable variation.
P a g e | 19
Control Charts
The Voice of the Process
Control chart A time-ordered plot of sample statistics, used to distinguish between random and
nonrandom variability.
An important tool in statistical process control is the control chart, which was developed by Walter
Shewhart. A control chart is a time-ordered plot of sample statistics. It is used to distinguish between
random variability and nonrandom variability. It has upper and lower limits, called control limits, that
define the range of acceptable (i.e., random) variation for the sample statistic.
The purpose of a control chart is to monitor process output to see if it is random. A necessary (but not
sufficient) condition for a process to be deemed “in control,” or stable, is for all the data points to fall
between the upper and lower control limits. Conversely, a data point that falls outside of either limit would
be taken as evidence that the process output may be nonrandom and, therefore, not “in control.” If that
happens, the process would be halted to find and correct the cause of the nonrandom variation. The essence
of statistical process control is to assure that the output of a process is random so that future output will be
random.
Mean Charts
A mean control chart, sometimes referred to as an x (“x -bar”) chart, is based on a normal distribution. It
can be constructed in one of two ways. The choice depends on what information is available. Although the
value of the standard deviation of a process, , is often unknown, if a reasonable estimate is
available, one can compute control limits using these formulas:
Upper control limit (UCL): = x + z x
were,
x= / n
x = Standard deviation of distribution of sample means
= Estimate of the process standard deviation
n = Sample size
z = Standard normal deviate
x = Average of sample means
The following example illustrates the use of these formulas.
SAMPLE
1 2 3 4 5
1 12.11 12.15 12.09 12.12 12.09
Observation 2 12.10 12.12 12.09 12.10 12.14
3 12.11 12.10 12.11 12.08 12.13
4 12.08 12.11 12.15 12.10 12.12
x 12.10 12.12 12.11 12.10 12.12
Solution
If the standard deviation of the process is unknown, another approach is to use the sample range as a
measure of process variability. The appropriate formulas for control limits are
Question:
Twenty samples of n=8 have been taken from a cleaning operation. The average sample range for the 20
samples was 0.016 minute, and the average mean was 3 minutes. Determine three sigma control limits for
this process.
Solution:
− −
−
𝑥 = 3 cm, 𝑅 = .016, A2=.37 for n=8.
− −
−
UCL = 𝑥 + 𝐴2 𝑅 = 3+.37(.016) = 3.006 minutes
− −
−
LCL = 𝑥 − 𝐴2 𝑅 = 3-.37(.016) = 2.994 minutes
Range Charts:
Range Control Chart(R-Charts) are used to monitor process dispersion; they are sensitive to changes in
process dispersion. Although the underlying sampling distribution is act normal. The concepts for use of
range charts are match the same as those for use of mean charts control limits for range charts are found
using the average sample range in conjunction with these formulas.
−
UCL=D4 𝑅
−
LCL=D3 𝑅
P a g e | 22
Number of Observation Factor for X Chart Lower Control Limit Upper Control Limit
in Subgroup
(A2) (D3) (D4)
(n)
2 1.88 0 3.27
3 1.02 0 2.57
4 0.73 0 2.28
5 0.58 0 2.11
6 0.48 0 2.00
If the process standard deviation is known, control limits for a range chart can be calculated using values
from this table.
P a g e | 23
Solution:
−
𝑅 = .01 𝑐𝑚, n=10
From Range Chart, for n =10, D4= 1.78 and D3= .22.
UCLR= 1.78(.01) = .0178 or .018
LCLR= .22(.01) = .0022 or .002.
P- chart
P a g e | 24
Use of p-chart
• When observation can be placed into one of two categories, example (good or bad)
• When the data consist of Multiple samples of N observation each (example- 15 sample of
n=20)
An inspector counted the number of defective monthly billing statements of a company telephone in each
of 20 samples. Using the following information, construct a control chart that will describe 99.74 percent
of the chance variation in the process when the process is in control. Each sample contained 100 statements.
Sample Number of Defectives Sample Number of Defectives
1 4 11 8
2 10 12 12
3 12 13 9
4 3 14 10
5 9 15 21
6 11 16 10
7 10 17 8
8 22 18 12
9 13 19 10
10 10 20 16
220
Solution:
Z for 99.74 percent is 3.00 (from Appendix Table A)
Total number of defectives 220
p = = =.11
Total number of obsevations 20(100)
σp= √𝑝(1−𝑝
𝑛
)
=√.11(1−.11)
100
=.03
Plotting the control limits and the sample fraction defective, you can see that the last value is above the
upper control limit. The process would be stopped at that point to find and correct the possible cause. Then
new data would be collected to establish new control limits. If no cause is found, this could be due to
chance. The new limits would remain but future output would be monitored to assure the process remains
in control.
C-chart
P a g e | 26
Use of C-chart
When only the Number of occurrences per unit of measure can be counted: nonoccurrence cannot be
counted be counted. Examples of occurrences and units of measure include.
a. Scratches, Chips, dents, or errors per item.
b. Cracks or faults per unit of distance (example: meters, miles).
c. Breaks or tears, per unit of area (example: square yard, square meter).
d. Bacteria or pollutants per unit of volume (example: gallon, cubic foot, cubic yard).
Rolls of coiled wire are monitored using a c -chart. Eighteen rolls have been examined, and the number of
defects per roll has been recorded in the following table. Is the process in control? Plot the values on a
control chart using three standard deviation control limits.
Sample Number of Defects Sample Number of Defects
1 3 10 1
2 2 11 3
3 4 12 4
4 5 13 2
5 1 14 4
6 2 15 2
7 4 16 1
8 1 17 3
9 2 18 1
45
Solution:
c = 45/18 = 2.5 = Average number of defects per coil
When the computed lower control limit is negative, the effective lower limit is zero. The calculation
sometimes produces a negative lower limit due to the use of the normal distribution to approximate the
Poisson distribution: The normal is symmetrical, whereas the Poisson is not symmetrical when c is close to
zero.
(Run Tests)
Control charts test for points that are too extreme to be considered random (e.g., points that are outside of
the control limits). However, even if all points are within the control limits, the data may still not reflect a
random process. In fact, any sort of pattern in the data would suggest a nonrandom process.
Analysts often supplement control charts with a run test, which checks for patterns in a sequence of
observations. This enables an analyst to do a better job of detecting abnormalities in a process and provides
insights into correcting a process that is out of control. A variety of run tests are available; this section
describes two that are widely used.
To determine whether any patterns are present in control chart data, one must transform the data into both
As and Bs and Us and Ds, and then count the number of runs in each case. These numbers must then be
compared with the number of runs that would be expected in a completely random series. For both the
median and the up/down run tests, the expected number of runs is a function of the number of observations
in the series. The formulas are
where N is the number of observations or data points, and E (r) is the expected number of runs. The actual
number of runs in any given set of observations will vary from the expected number, due to chance and any
patterns that might be present. Chance variability is measured by the standard deviation of runs. The
formulas are:
Distinguishing chance variability from patterns requires use of the sampling distributions for median runs
and up/down runs. Both distributions are approximately normal.
Z test
P a g e | 29
Mathematical Representation on the basis of Run test: Twenty sample means have been taken from a
process. The means are shown in the following table. Use median and up\down run tests with z=2 to
determine if assignable causes of variation are present. Assume the medium is 11.0.
The means are marked according to above\below the median and up\down. The solid lines represent the
runs.
Sample A\B Mean U\D Sample A\B Mean U\D
1 |B !0.0 ---- 11 |B 10.7 |D
2 |B 10.4 |U 12 |A 11.3 |U
3 |B 10.2 |D 13 |B 10.8 |D
4 |A 11.5 |U 14 |A 11.8 |U
5 |B 10.8 |D 15 |A 11.2 |D
6 |A 11.6 |U 16 |A 11.6 |U
7 |A 11.1 |D 17 |A 11.2 |D
8 |A 11.2 |U 18 |B 10.6 |D
9 |B 10.6 |D 19 |B 10.7 |U
10 |B 10.9 |U 20 |A 11.9 |U
A\B:10 runs U\D:17 runs
𝑬 𝟐𝑵−𝟏 𝟐(𝟐𝟎)−𝟏
(𝒓)𝑼∕𝒅 = = =𝟏𝟑
𝟑 𝟑
𝑵−𝟏 𝟐𝟎 − 𝟏
𝝈𝒎𝒆𝒅 = √ =√ = 𝟐 ⋅ 𝟏𝟖
𝟒 𝟒
𝟏𝟔𝑵 − 𝟐𝟗 𝟏𝟔(𝟐𝟎) − 𝟐𝟗
𝝈𝑼∕𝒅 = √ =√ = 𝟏 ⋅ 𝟖𝟎
𝟗𝟎 𝟗𝟎
Control limits
The dividing lines between random and nonrandom deviations from the mean of the distribution.
Type I error:
Concluding a process is not in control when it actually is.
Type II error:
Concluding a process is in control when it is not.
Mathematical Representation
Type I error (alpha risk). After several investigations of points outside control limits revealed nothing, a
manager began to wonder about the probability of a Type I error for the control limits used (z=1.90).
Solution:
a. Using Appendix B, Table A, find that the area under the curve between z = 0 and z =+1.90 is .4713.
Therefore, the area (probability) of values within -1.90 to +1.90 is 2(.4713) = .9426, and the area beyond
these values is 1 − .9426 = .0574. Hence, the alpha risk is 5.74 percent.
b. The alpha risk (Type I error probability) is always specified as an area in the tail(s) of a distribution.
With control charts, you use two-sided control limits. Consequently, half of the risk lies in each tail. Hence,
the area in the right tail is 1 percent, or .0100. This means that .4900 should be the area under the curve
between z = 0 and the value of z you are looking for. The closest value is .4901 for z = 2.33. Thus, control
limits based on z =+2.33 provide an alpha risk of about 2 percent.
P a g e | 31
Reference:
1: operation management by William j. Stevenson, Rochester institute of technology. 8th edition.
2: operation management by William j. Stevenson, Rochester institute of technology. 11th edition.