Sie sind auf Seite 1von 21

Cpk Vs.

ppm Table
The performance of a process may be characterized in terms of how close it gets to hitting
its target or meeting its specifications and how consistent it is in doing so. For a process
whose output data comprise a normal distribution, its performance can be conveniently
quantified in terms of its process capability index, Cpk.
The Cpk of a process measures how centered the output of the process is between its
lower and upper specification limits and how variable (and therefore how stable or nonstable) the output is. In fact, the Cpk is expressed as the ratio of how far the mean of the
output data is from the closer spec limit (the centering of the process) to three times their
standard deviation (the process variability).
If the mean of the process data is closer to the lower spec limit LSL and the standard
deviation of the process data is Stdev, then Cpk = (Mean-LSL) / (3 Stdev). If the mean of
the process data is closer to the upper spec limit USL, then
Cpk = (USL-Mean) / (3 Stdev).
An ideal process is one whose output is always dead center between the spec limits, such
that the mean of its output data equals this dead center and the standard deviation is zero.
The Cpk of this ideal process is infinite.
As a process becomes less centered between the spec limits or as it becomes more
variable, its Cpk decreases. As its Cpk decreases, the probability of it exhibiting an output
that is outside its specification limits increases. Thus, every Cpk value corresponds to a
percent defective rate, which may be expressed in parts per million, or ppm.
Table 1 shows some Cpk values and their equivalent ppm rates. In the semiconductor
industry, the Cpk goal for a process is normally set at 1.67, although a Cpk of 1.33 is still
considered acceptable.
Table 1. Cpk Vs. ppm
Cpk

Sigma

1-sided Spec ppm

2-sided Spec ppm

0.43
0.47
0.50
0.53
0.57
0.60
0.63
0.67
0.70
0.73
0.77
0.80
0.83
0.87
0.90
0.93
0.97

1.50
-

96,800
80,755
66,805
54,800
44,565
35,980
28,715
22,750
17,865
13,905
10,725
8,200
6,210
4,661
3,467
2,555
1,866

193,600
161,510
133,610
109,600
89,130
71,960
57,430
45,500
35,730
27,810
21,450
16,400
12,420
9,322
6,934
5,110
3,732

1.00
1.03
1.07
1.10
1.13
1.16
1.20
1.23
1.27
1.30
1.33
1.37
1.40
1.43
1.47
1.50
1.53
1.57
1.60
2.00

3.00
~3.50
4.00
4.50
~5.00
6.00

1350
967
687
483
337
232
159
108
73
49
32
20.5
13.5
8.5
5.5
3.5
2
1.5
0.25
0.00099

2,700
1,935
1.374
967
674
465
318
216
145
98
64
41
27
17
11
7
4
3
0.5
0.00198

Statistical Process Control (SPC)


Statistical Process Control (SPC) is a system for monitoring, controlling, and improving a
process through statistical analysis. It has many aspects, from control charting to process
capability studies and improvement. Nonetheless, the over-all SPC system of a company
may be broken down into four basic steps: 1) measuring the process; 2) eliminating
variances within the process to make it consistent; 3) monitoring the process; and 4)
improving the process. This four-step cycle may be employed over and over again for
continuous improvement.
Bulk of SPC concepts in use today were developed based on the premise that the process
parameter being controlled follows a normal distribution. Any SPC practitioner must be
aware that the parameter must first be confirmed to be normal before being subjected to
analysis concepts based on normal behavior. Thus, any discussion on SPC must be
preceded by a discussion of what a normal distribution is.

LINKS:
Control Charting
Process
Capability
Cpk Vs. ppm
Table
Monitors &
Controls
Quality Systems

The Normal Distribution


The normal distribution (see Fig. 1), normal curve, or bell-shaped curve, is probably the most
recognized and most widely-used statistical distribution. The reason for this is that many
physical, biological, and social parameters obey the normal distribution. Such parameters are
then said to behave 'normally' or, more simply, are said to be 'normal.' The semiconductor
industry has many processes that output data or results that comprise a normal distribution.
As such, it is important for every process engineer to have a firm grasp of what a normal
distribution is.
Aside from the fact that the normal distribution is frequently encountered in our day-to-day
lives, the mathematics governing normal behavior are fairly simple. In fact, only two
parameters are needed to describe a normal distribution, namely, the mean or its center, and
the standard deviation (also known as sigma) or its variability. Knowing both parameters is
equivalent to knowing how the distribution looks like.

Figure 1. The Normal Distribution

The normal distribution is bell-shaped, i.e., it peaks at the center and tapers off outwardly
while remaining symmetrical with respect to the center. To illustrate this in more tangible
terms, imagine taking down the height of every student in a randomly selected Grade 5
class and plotting the measurements on a chart whose x-axis corresponds to the height of
the student and whose y-axis corresponds to the number of students.
What is expected to emerge from this exercise is a normal curve, wherein a big slice of the
student population will have a height that is somewhere in the middle of the distribution,
say 57-59 inches tall. The number of students belonging to other height groups will be
less than the number of students in the 57"-59" category .
In fact, the number of students decreases at a calculable rate as the height group moves
further away from the center. Eventually you might find only one shortest student at, say,
48", and one tallest student who probably stands at 66". Lastly, plotting the number of the
students falling under different height ranges of equal intervals will result in a bell-shaped
curve. Such a plot is called a histogram, a simple example of which is shown in Figure 2.

Figure 2. Example of a histogram of heights of students in a Grade 5


class; the y-axis corresponds to the number of students per category

Statistical Process Control (SPC) (Page 2 of 4)


What's notable about normal distributions is that regardless of their standard deviation
value, the % of data falling under a given number of standard deviations is constant. For
example, say that the standard deviation of process 1 is 100, and the standard deviation of
process 2 is 200. Process 1 and Process 2 will have different data distribution shapes
(Process 1 being more stable), but for both processes, 66% of the data under the normal
curve will fall within +/- one (1) standard deviation from the mean of the distribution (i.e.,
between {mean - 1 sigma} and {mean + 1 sigma}), and 37% of the data will be outside it.
Table 1 shows the percentages of data falling under different numbers of sigma.

LINKS:
Control Charting
Process
Capability
Cpk Vs. ppm
Table
Monitors &
Controls
Quality Systems

Table 1. % Data Falling Under Different Numbers of +/- Sigma

# of Sigma's

% of Data
Covered

% of Data
Outside

+/- 1 Sigma

66%

37%

+/- 2 Sigmas

95%

5%

+/- 3 Sigmas

99.73%

0.27%

+/- 4 Sigmas

99.9936%

0.0063%

+/- 5 Sigmas

99.99995%

0.00005%

Skewed Distributions
Perfectly normal curves are hard to come by with finite samples or data. Thus, some data
distributions that are theoretically normal may not appear to be one once the data are
plotted, i.e., the mean may not be at the center of the distribution or there may be slight
non-symmetry. If a normal distribution appears to be 'heavy' or leaning towards the right
side of the distribution, it is said to be skewed to the left. A normal distribution that's
leaning to the left is said to be skewed to the right.
Many response parameters encountered in the semiconductor industry behave normally,
which is why statistical process control has found its way extensively into this industry.
The objective of SPC is to produce data distributions that are stable, predictable, and well
within the specified limits for the parameter being controlled.
In relation to the preceding discussions, this is equivalent to achieving data distributions
that are centered between the specified limits, and as narrow as possible. Good centering
between limits and negligible variation translates to parameters that are always within
specifications, which is the true essence of process control.
Control Charting
It is often said that you can not control something that you do not measure. Thus, every
engineer setting up a new process must have a clear idea of how the performance of this
new process is to be measured. Since every process needs to satisfy customer
requirements, process output parameters for measurement and monitoring are generally
based on customer specifications. Industry-accepted specifications are also followed in
selecting process parameters for monitoring.
Control charting is a widely-used tool for process monitoring in the semiconductor
industry. It employs control charts (see Fig. 3), which are simply plots of the process
output data over time. Before a control chart may be used, the process engineer must first
ensure that the process to be monitored is normal and stable.
A process may have several control charts - one for each of its major output parameters. A
new control chart must have at least the following: the properly labeled x- and y-axes, lines
showing the lower and upper specification limits for the parameter being monitored, and a
line showing the center or target of these specifications. Once a control chart has been set
up, the operator must diligently plot the output data at predefined intervals.
After 30 data points have been collected on the chart (may be less if measurement intervals
are long), the upper and lower control limits of the process may already be computed.
Control limits define the boundaries of the normal behavior of the process. Their values
depend only on the output data generated by the process in the immediate past. Control
limits are therefore independent of specification limits. However, both sets of limits are used
in the practice of SPC, although in different ways.

Statistical Process Control (SPC) (Page 3 of 4)


The lower control limit LCL and the upper control limit UCL of a process may be calculated
from the mean and standard deviation (or sigma) of the plotted data as follows:

LCL = Mean - (3 x Sigma);


UCL = Mean + (3 x Sigma).
Thus, the span from the LCL of a process to its UCL is 6 sigma. The probability of getting
points outside this +/- 3 sigma range is already very low (see Table 1). Getting a
measurement outside this range should therefore warn an engineer that something
abnormal is happening, i.e., the process may be going out of control. This is the reason
why these boundaries are known as 'control limits.'
LINKS:
Control Charting
Process
Capability
Cpk Vs. ppm
Table
Monitors &
Controls
Quality Systems

Once the control limits have been included on the control charts (also in the form of
horizontal lines like the specification limits), the operator can start using the chart visually
to detect anomalous trends in the process that she would need to notify the engineer
about.

Figure 3. Example of a control chart showing data that are


slightly off-centered, but nonetheless in control and within specs

For instance, any measurement outside the control limits is an automatic cause for alarm,
because the probability of getting such a measurement is low. Four (4) or more
consecutively increasing or decreasing points form a trend that is not normal, and
therefore deserves attention. Six (6) consecutive points on one side of the mean also
deserve investigation. When such abnormalities are observed, the process owner must
take an action to bring the process back to its normal behavior.
Control limits must be recomputed regularly (say, every quarter), to ensure that the control
limits being used by the operator are reflective of the current process behavior.

Read more about: Control Charting.


The Process Capability Indices
Being able to monitor a process for out-of-control situations is one thing; knowing how a
process actually performs is another. Eyeballing the centering and shape of a data
distribution can give us quick, useful information on how the corresponding process
behaves, but it is not very helpful in quantifying the process' actual and potential
performance. It is for this reason that statisticians have come up with methods for
expressing the behavior or capability of process distributions in terms of single numbers
known as process capability indices.
Process capability refers to the ability of a process to meet customer requirements or
specification limits, i.e., how consistent its output is in being within its lower and upper spec
limits. A process capability index should therefore be able to indicate how well the process
can meet its specs.
The most basic process capability index is known as the simple process capability index,
denoted by 'Cp'. Cp quantifies the stability of a process, i.e., the consistency of its output.
As mentioned earlier, the process capability indices discussed here presume the normality
of the process. As such, the inconsistency of the process may be measured in terms of
the standard deviation or sigma of the output data of the process. This is what Cp does - it
uses the sigma to quantify the variation of a process, and compares it against the distance
between the upper spec limit (USL) and lower spec limit (LSL) of the process. In
mathematical form:
Cp = (USL - LSL) / (6 x Sigma)
The quantity (USL - LSL) is basically the range of output that the process must meet, while
6 sigma corresponds to +/- 3 sigma from the mean, or 99.73% of all the process output
data. The smaller the value of 6 sigma, the narrower the process output distribution is,
denoting higher stability. Thus, Cp increases as process stability increases. Thus, a
process needs a Cp > 1 to ensure that it is narrow enough to meet the spec range 99.73%
of the time

Statistical Process Control (SPC) (Page 4 of 4)


Although Cp indicates the stability of a process, it has one major drawback that makes it
almost useless in the semiconductor industry. It does not consider the centering of the
process distribution within the spec limits. A process with a Cp of 100 may be very stable,
with all its output data very close to each other, but it may also be out-of-spec at all times,
i.e., if it is centered outside the spec limits!

LINKS:
Control Charting
Process
Capability
Cpk Vs. ppm
Table
Monitors &
Controls
Quality Systems

This weakness of Cp is addressed by another process capability index, Cpk. Cpk


measures how centered the output of the process is between its lower and upper limits, as
well as how variable the output is. Cpk is expressed as the ratio of how far the mean of
the output data is from the closer spec limit (the centering of the process) to three times
their standard deviation (the process variability).
CPL = (mean - LSL) / (3 sigma) : process capability index for single-sided (lower) spec limit
CPU = (USL - mean) / (3 sigma) : process capability index for single-sided (upper) spec
limit
Cpk = min{CPL,CPL} : process capability index for two-sided spec limits
What these formulae mean is this: Cpk is equal to whichever is lower between CPL and
CPU. If the mean of the process data is closer to the lower spec limit LSL, then Cpk = C PL.
If the mean of the process data is closer to the upper spec limit USL, then Cpk = C PU.
An ideal process is one whose output is always dead center between the spec limits, such
that the mean of its output data equals this dead center and the standard deviation is
zero. The Cpk of this ideal process is infinite (so is the Cpk of other processes whose
sigma = 0, as long as the LSL<mean<USL).
The Cpk decreases if one or both of the following occurs: 1) the data become less
centered; and 2) the data become more variable (sigma increases). Thus, improving the
process capability of a process entails one or both of: 1) centering the output between
limits and 2) decreasing the variation of the output data.
The essence of SPC, therefore, is being able to recognize whether a low Cpk is due to the
mean of the process or its sigma, and taking the necessary actions to correct the problem,
be it centering of the data or making them less variable. In any process, the actions
needed to center the output data may be different from what needs to be done to make the
data less variable. Knowledge of this basic SPC principle is therefore a necessary weapon
in every process engineer's arsenal.
As of this writing, most semiconductor companies target a Cpk of 1.67 for their processes,
although they would be satisfied to have an actual Cpk of at least 1.33. Everything, of
course, depends on what spec limits the customer imposes on the manufacturer. Still, at
the end of the day it should always be the manufacturer's goal to center their output
between these spec limits as consistently as possible.

===================

Six Sigma vs. Three Sigma


Six Sigma vs. Three Sigma
1.5 Sigma Process Shift Explanationi

SixSigma recently released a process sigma calculator which allows the operator to
input process opportunities and defects and easily calculate the process sigma to
determine how close (or far) a process is from 6 sigma. One of the caveats written in fine
print refers to the calculator using a default process shift of 1.5 sigma. From an earlier
poll, greater than 50% of polled quality professionals indicated that they are not aware of
why a process may shift 1.5 sigma. My goal is to explain it here.
I'm not going to bore you with the hard core statistics. There's a whole statistical section
dealing with this issue, and every green, black and master black belt learns the calculation
process in class. If you didn't go to class (or you forgot!), the table of the standard normal
distribution is used in calculating the process sigma. Most of these tables, however, end at
a z value of about 3 (see the iSixSigma table for an example). In 1992, Motorola
published a book (see chapter 6) entitled Six Sigma Producibility Analysis and Process
Characterizationbuy it now!, written by Mikel J. Harry and J. Ronald Lawson. In it is one of
the only tables showing the standard normal distribution table out to a z value of 6.

Using this table you'll find that 6 sigma actually translates to about 2 defects per billion
opportunities, and 3.4 defects per million opportunities, which we normally define as 6
sigma, really corresponds to a sigma value of 4.5. Where does this 1.5 sigma difference
come from? Motorola has determined, through years of process and data collection, that
processes vary and drift over time - what they call the Long-Term Dynamic Mean
Variation. This variation typically falls between 1.4 and 1.6.
After a process has been improved using the Six Sigma DMAIC methodology, we
calculate the process standard deviation and sigma value. These are considered to be
short-term values because the data only contains common cause variation -- DMAIC
projects and the associated collection of process data occur over a period of months,
rather than years. Long-term data, on the other hand, contains common cause variation
and special (or assignable) cause variation. Because short-term data does not contain this
special cause variation, it will typically be of a higher process capability than the longterm data. This difference is the 1.5 sigma shift. Given adequate process data, you can
determine the factor most appropriate for your process.
In Six Sigma, The Breakthrough Management Strategy Revolutionizing The World's Top
Corporations, Harry and Schroeder write:
"By offsetting normal distribution by a 1.5 standard deviation on either side, the adjustment takes
into account what happens to every process over many cycles of manufacturing Simply put,
accommodating shift and drift is our 'fudge factor,' or a way to allow for unexpected errors or
movement over time. Using 1.5 sigma as a standard deviation gives us a strong advantage in
improving quality not only in industrial process and designs, but in commercial processes as well.
It allows us to design products and services that are relatively impervious, or 'robust,' to natural,
unavoidable sources of variation in processes, components, and materials."

Statistical Take Away: The reporting convention of Six Sigma requires the process
capability to be reported in short-term sigma -- without the presence of special cause
variation. Long-term sigma is determined by subtracting 1.5 sigma from our short-term
sigma calculation to account for the process shift that is known to occur over time.
This topic was revisited and more information has been provided.
================
Measuring Your Process Capability

Featured
Article

- Dr. Mehernosh Kapadia, GM, Quality Engineering, Supply Chain Management


Business Group, Tata AutoComp Systems Limited.
Preamble :
This article is devoted to the topic of process
capability, with the objective of making people
aware of this subject and its significance to
business success.
The author believes that personal awareness is a
prerequisite to personal action, and personal
action is what we need for success.
It can be a source material for you to use in

discussing this topic with your organization.


It will address issues like what is process capability,
how to measure it, and how to calculate the process
capability indices (Cp, Cpk).
It will also attempt to explain the differences between
process capability and process performance;
relationship between Cpk and non-conforming
(defect) rate; and illustrate the four outcomes of
comparing natural process variability with customer
specifications.

Send this article to colleagues and


friends
Send us your comments
Download this article in PDF Format

Related software from Symphony


Freeware Process Capability
Calculator
Non-Normal Process Capability
Evaluator

Lastly a commentary is provided on precautions we


should take while conducting process capability
studies.

What is Process Capability ?


1. Process capability is the long-term performance level of the process after it
has been brought under statistical control. In other words, process
capability is the range over which the natural variation of the process
occurs as determined by the system of common causes.
2. Process capability is also the ability of the combination of people, machine,
methods, material, and measurements to produce a product that will
consistently meet the design requirements or customer expectation.

Bombay
He can be contacted at e-mail address:
mkapadia@tacogroup.com
or through us at webmaster@symphonytech.com

Das könnte Ihnen auch gefallen