The process capability index, or process capability ratio, is a statistical measure of process capability: the ability of an engineering process to produce an output within specification limits. [1] The concept of process capability only holds meaning for processes that are in a state of statistical control. This means it cannot account for deviations which are not expected, such as misaligned, damaged, or worn equipment. Process capability indices measure how much "natural variation" a process experiences relative to its specification limits, and allows different processes to be compared to how well an organization controls them. Somewhat counterintuitively, higher index values indicate better performance, with zero indicating high deviation.
A company produces axles with nominal diameter 20 mm on a lathe. As no axle can be made to exactly 20 mm, the designer specifies the maximum admissible deviations (called tolerances or specification limits). For instance, the requirement could be that axles need to be between 19.9 and 20.2 mm. The process capability index is a measure for how likely it is that a produced axle satisfies this requirement. The index pertains to statistical (natural) variations only. These are variations that naturally occur without a specific cause. Errors not addressed include operator errors, or play in the lathe's mechanisms resulting in a wrong or unpredictable tool position. If errors of the latter kinds occur, the process is not in a state of statistical control. When this is the case, the process capability index is meaningless.
If the upper and lower specification limits of the process are USL and LSL, the target process mean is T, the estimated mean of the process is and the estimated variability of the process (expressed as a standard deviation) is , then commonly accepted process capability indices include:
Index | Description |
---|---|
Estimates what the process is capable of producing if the process mean were to be centered between the specification limits. Assumes process output is approximately normally distributed. | |
Estimates process capability for specifications that consist of a lower limit only (for example, strength). Assumes process output is approximately normally distributed. | |
Estimates process capability for specifications that consist of an upper limit only (for example, concentration). Assumes process output is approximately normally distributed. | |
Estimates what the process is capable of producing, considering that the process mean may not be centered between the specification limits. (If the process mean is not centered, overestimates process capability.) if the process mean falls outside of the specification limits. Assumes process output is approximately normally distributed. | |
Estimates process capability around a target, T. is always greater than zero. Assumes process output is approximately normally distributed. is also known as the Taguchi capability index. [2] | |
Estimates process capability around a target, T, and accounts for an off-center process mean. Assumes process output is approximately normally distributed. |
is estimated using the sample standard deviation.
Process capability indices are constructed to express more desirable capability with increasingly higher values. Values near or below zero indicate processes operating off target ( far from T) or with high variation.
Fixing values for minimum "acceptable" process capability targets is a matter of personal opinion, and what consensus exists varies by industry, facility, and the process under consideration. For example, in the automotive industry, the Automotive Industry Action Group sets forth guidelines in the Production Part Approval Process, 4th edition for recommended Cpk minimum values for critical-to-quality process characteristics. However, these criteria are debatable and several processes may not be evaluated for capability just because they have not properly been assessed.
Since the process capability is a function of the specification, the Process Capability Index is only as good as the specification. For instance, if the specification came from an engineering guideline without considering the function and criticality of the part, a discussion around process capability is useless, and would have more benefits if focused on what are the real risks of having a part borderline out of specification. The loss function of Taguchi better illustrates this concept.
At least one academic expert recommends [3] the following:
Situation | Recommended minimum process capability for two-sided specifications | Recommended minimum process capability for one-sided specification |
---|---|---|
Existing process | 1.33 | 1.25 |
New process | 1.50 | 1.45 |
Safety or critical parameter for existing process | 1.50 | 1.45 |
Safety or critical parameter for new process | 1.67 | 1.60 |
Six Sigma quality process | 2.00 | 2.00 |
However where a process produces a characteristic with a capability index greater than 2.5, the unnecessary precision may be expensive. [4]
The mapping from process capability indices, such as Cpk, to measures of process fallout is straightforward. Process fallout quantifies how many defects a process produces and is measured by DPMO or PPM. Process yield is the complement of process fallout and is approximately equal to the area under the probability density function if the process output is approximately normally distributed.
In the short term ("short sigma"), the relationships are:
Cp | Sigma level (σ) | Area under the | Process yield | Process fallout (in terms of DPMO/PPM) |
---|---|---|---|---|
0.33 | 1 | 0.6826894921 | 68.27% | 317311 |
0.67 | 2 | 0.9544997361 | 95.45% | 45500 |
1.00 | 3 | 0.9973002039 | 99.73% | 2700 |
1.33 | 4 | 0.9999366575 | 99.99% | 63 |
1.67 | 5 | 0.9999994267 | 99.9999% | 1 |
2.00 | 6 | 0.9999999980 | 99.9999998% | 0.002 |
In the long term, processes can shift or drift significantly (most control charts are only sensitive to changes of 1.5σ or greater in process output). If there was a 1.5 sigma shift 1.5σ off of target in the processes (see Six Sigma), it would then produce these relationships: [5]
Cp | Adjusted Sigma level (σ) | Area under the | Process yield | Process fallout (in terms of DPMO/PPM) |
---|---|---|---|---|
0.33 | 1 | 0.3085375387 | 30.85% | 691462 |
0.67 | 2 | 0.6914624613 | 69.15% | 308538 |
1.00 | 3 | 0.9331927987 | 93.32% | 66807 |
1.33 | 4 | 0.9937903347 | 99.38% | 6209 |
1.67 | 5 | 0.9997673709 | 99.9767% | 232.6 |
2.00 | 6 | 0.9999966023 | 99.99966% | 3.40 |
Because processes can shift or drift significantly long term, each process would have a unique sigma shift value, thus process capability indices are less applicable as they require statistical control.
Consider a quality characteristic with target of 100.00 μm and upper and lower specification limits of 106.00 μm and 94.00 μm respectively. If, after carefully monitoring the process for a while, it appears that the process is in control and producing output predictably (as depicted in the run chart below), we can meaningfully estimate its mean and standard deviation.
If and are estimated to be 98.94 μm and 1.03 μm, respectively, then
Index |
---|
The fact that the process is running off-center (about 1σ below its target) is reflected in the markedly different values for Cp, Cpk, Cpm, and Cpkm.
In probability theory and statistics, a normal distribution or Gaussian distribution is a type of continuous probability distribution for a real-valued random variable. The general form of its probability density function is The parameter is the mean or expectation of the distribution, while the parameter is the variance. The standard deviation of the distribution is (sigma). A random variable with a Gaussian distribution is said to be normally distributed, and is called a normal deviate.
In statistics, the standard deviation is a measure of the amount of variation of the values of a variable about its mean. A low standard deviation indicates that the values tend to be close to the mean of the set, while a high standard deviation indicates that the values are spread out over a wider range. The standard deviation is commonly used in the determination of what constitutes an outlier and what does not. Standard deviation may be abbreviated SD or Std Dev, and is most commonly represented in mathematical texts and equations by the lowercase Greek letter σ (sigma), for the population standard deviation, or the Latin letter s, for the sample standard deviation.
In probability theory, a log-normal (or lognormal) distribution is a continuous probability distribution of a random variable whose logarithm is normally distributed. Thus, if the random variable X is log-normally distributed, then Y = ln(X) has a normal distribution. Equivalently, if Y has a normal distribution, then the exponential function of Y, X = exp(Y), has a log-normal distribution. A random variable which is log-normally distributed takes only positive real values. It is a convenient and useful model for measurements in exact and engineering sciences, as well as medicine, economics and other topics (e.g., energies, concentrations, lengths, prices of financial instruments, and other metrics).
In statistics, the standard score is the number of standard deviations by which the value of a raw score is above or below the mean value of what is being observed or measured. Raw scores above the mean have positive standard scores, while those below the mean have negative standard scores.
Six Sigma (6σ) is a set of techniques and tools for process improvement. It was introduced by American engineer Bill Smith while working at Motorola in 1986.
A Z-test is any statistical test for which the distribution of the test statistic under the null hypothesis can be approximated by a normal distribution. Z-test tests the mean of a distribution. For each significance level in the confidence interval, the Z-test has a single critical value which makes it more convenient than the Student's t-test whose critical values are defined by the sample size. Both the Z-test and Student's t-test have similarities in that they both help determine the significance of a set of data. However, the z-test is rarely used in practice because the population deviation is difficult to determine.
Control charts are graphical plots used in production control to determine whether quality and manufacturing processes are being controlled under stable conditions. The hourly status is arranged on the graph, and the occurrence of abnormalities is judged based on the presence of data that differs from the conventional trend or deviates from the control limit line. Control charts are classified into Shewhart individuals control chart and CUSUM(CUsUM)(or cumulative sum control chart)(ISO 7870-4).
In statistics, an effect size is a value measuring the strength of the relationship between two variables in a population, or a sample-based estimate of that quantity. It can refer to the value of a statistic calculated from a sample of data, the value of one parameter for a hypothetical population, or to the equation that operationalizes how statistics or parameters lead to the effect size value. Examples of effect sizes include the correlation between two variables, the regression coefficient in a regression, the mean difference, or the risk of a particular event happening. Effect sizes are a complement tool for statistical hypothesis testing, and play an important role in power analyses to assess the sample size required for new experiments. Effect size are fundamental in meta-analyses which aim to provide the combined effect size based on data from multiple studies. The cluster of data-analysis methods concerning effect sizes is referred to as estimation statistics.
In statistical inference, specifically predictive inference, a prediction interval is an estimate of an interval in which a future observation will fall, with a certain probability, given what has already been observed. Prediction intervals are often used in regression analysis.
In probability theory and statistics, the coefficient of variation (CV), also known as normalized root-mean-square deviation (NRMSD), percent RMS, and relative standard deviation (RSD), is a standardized measure of dispersion of a probability distribution or frequency distribution. It is defined as the ratio of the standard deviation to the mean , and often expressed as a percentage ("%RSD"). The CV or RSD is widely used in analytical chemistry to express the precision and repeatability of an assay. It is also commonly used in fields such as engineering or physics when doing quality assurance studies and ANOVA gauge R&R, by economists and investors in economic models, and in psychology/neuroscience.
In statistics, the mode is the value that appears most often in a set of data values. If X is a discrete random variable, the mode is the value x at which the probability mass function takes its maximum value. In other words, it is the value that is most likely to be sampled.
The process capability is a measurable property of a process to the specification, expressed as a process capability index or as a process performance index. The output of this measurement is often illustrated by a histogram and calculations that predict how many parts will be produced out of specification (OOS).
The Z-factor is a measure of statistical effect size. It has been proposed for use in high-throughput screening (HTS), where it is also known as Z-prime, to judge whether the response in a particular assay is large enough to warrant further attention.
In statistics, the 68–95–99.7 rule, also known as the empirical rule, and sometimes abbreviated 3sr, is a shorthand used to remember the percentage of values that lie within an interval estimate in a normal distribution: approximately 68%, 95%, and 99.7% of the values lie within one, two, and three standard deviations of the mean, respectively.
In statistics and in particular statistical theory, unbiased estimation of a standard deviation is the calculation from a statistical sample of an estimated value of the standard deviation of a population of values, in such a way that the expected value of the calculation equals the true value. Except in some important situations, outlined later, the task has little relevance to applications of statistics since its need is avoided by standard procedures, such as the use of significance tests and confidence intervals, or by using Bayesian analysis.
In process improvement efforts, the process performance index is an estimate of the process capability of a process during its initial set-up, before it has been brought into a state of statistical control.
Process window index (PWI) is a statistical measure that quantifies the robustness of a manufacturing process, e.g. one which involves heating and cooling, known as a thermal process. In manufacturing industry, PWI values are used to calibrate the heating and cooling of soldering jobs while baked in a reflow oven.
In probability theory and statistics, the index of dispersion, dispersion index, coefficient of dispersion, relative variance, or variance-to-mean ratio (VMR), like the coefficient of variation, is a normalized measure of the dispersion of a probability distribution: it is a measure used to quantify whether a set of observed occurrences are clustered or dispersed compared to a standard statistical model.
Experimental uncertainty analysis is a technique that analyses a derived quantity, based on the uncertainties in the experimentally measured quantities that are used in some form of mathematical relationship ("model") to calculate that derived quantity. The model used to convert the measurements into the derived quantity is usually based on fundamental principles of a science or engineering discipline.
In statistics, the strictly standardized mean difference (SSMD) is a measure of effect size. It is the mean divided by the standard deviation of a difference between two random values each from one of two groups. It was initially proposed for quality control and hit selection in high-throughput screening (HTS) and has become a statistical parameter measuring effect sizes for the comparison of any two groups with random values.
{{cite web}}
: External link in |work=
(help)