Sunteți pe pagina 1din 7

An Assessment of RPN Prioritization in a

Failure Modes Effects and Criticality Analysis


John B. Bowles • University of South Carolina • Columbia
Key Words: Risk Priority Number, FMEA, FMECA

SUMMARY & CONCLUSIONS Ground Vehicle Recommended Practice J1739 [3], provides a
The Risk Priority Number methodology for prioritizing very detailed and comprehensive methodology for performing
failure modes is an integral part of the Automobile FMECA a Failure Modes Effects and Criticality Analysis (FMECA).
technique. The technique consists of ranking the potential Since the major automobile companies and their suppliers
failures from 1 to 10 with respect to their severity, probability constitute a large part of the US economy the procedure has
of occurrence, and likelihood of detection in later tests, and become very widespread. It is also very well documented and
multiplying the numbers together. The result is a numerical serves as a model for other industries. The Ford procedure
ranking, called the RPN, on a scale from 1 to 1000. Potential was one of the first to extend the FMEA methodology to
failure modes having higher RPNs are assumed to have a potential manufacturing process induced failures as well as
higher design risk than those having lower numbers. Although design induced failures.
it is well documented and easy to apply, the method is An integral part of the Automobile FMEA technique is the
seriously flawed from a technical perspective. This makes the Risk Priority Number (RPN) methodology for ranking and
interpretation of the analysis results problematic. The assessing the design risk of potential failure modes. The RPN
problems with the methodology include the use of the ordinal is developed by assigning potential failure modes a rank from
ranking numbers as numeric quantities, the presence of holes 1 to 10 with respect to the severity of the failure mode effect
making up a large part of the RPN measurement scale, (S), its probability of occurrence (O), and the likelihood of its
duplicate RPN values with very different characteristics, and being detected in later design evaluation tests (D). These
varying sensitivity to small changes. Recommendations for an ranking factors are then multiplied to give the RPN:
improved methodology are also given. RPN = S × O × D
Since the three ranking factors range from 1 to 10, the
1. INTRODUCTION
resulting RPN ranges from 1 to 1000. Higher RPNs are
"If you cannot measure it, you cannot improve it." assumed to be more of a design risk than those having a lower
"When you measure what you are speaking about and RPN [3]. Also, multiplying the S, O, and D factors magnifies
express it in numbers, you know something about it, but when the effect of any high risk factors. The RPN number is used to
you cannot express it in numbers your knowledge about it is of prioritize failure modes for corrective actions and to evaluate
a meager and unsatisfactory kind." the improvement (change in RPN) resulting from prospective
— Sir William Thomson, Lord Kelvin (1824 – 1907) [1] design changes. Corrective actions should be aimed at
reducing one of the S, O, or D rankings.
As summarized by Lord Kelvin, in any engineering or From a management perspective the RPN technique is
scientific endeavor, it is necessary to be able to measure and simple, easy to understand, straight forward to use, well
quantify the attributes of an item if we wish to model, documented, and with appropriate training, consistent results
understand, and predict the behavior of the item in question. can be obtained. Unfortunately, from a technical perspective
In a Failure Modes and Effects Analysis (FMEA) the the method is seriously flawed which makes both the analysis
objective is to determine how an item can fail and assess the and the interpretation of the results problematic. The
effect of that failure on the system of which the item is a part; implications of these flaws are the focus of this paper.
the term, “system”, in this case is used to mean not only the We begin in Section 2 with a discussion of measurement
specific, end-item piece of equipment, but also the users, methodologies. Section 3 examines each of the RPN scales:
maintainers, and owners of the equipment. The results of the Severity, Occurrence, and Detection, with respect to the types
analysis help managers and engineers to determine which of measurement each represents. Section 4 then examines the
failure modes pose a hazard that must be designed out of the formation of the RPN number itself and some of the
product, which ones can be handled by appropriate corrective mathematical implications from its construction. Conclusions
actions and mitigation procedures, and which ones can be and recommendations are in Section 5.
safely ignored. Quantifying the effects of the failure with
respect to the severity of the effect and the failure mode’s 1.1. Acronyms and Notation
probability of occurrence helps in this regard. RPN Risk Priority Number
The Automobile Failure Modes and Effects Analysis S Severity ranking (1 to 10)
technique, originally developed at Ford Motor Company [2] O Occurrence Ranking (1 to 10)
and later standardized for the automobile industry as SAE D Detection Ranking (1 to 10)

0-7803-7717-6/03/$17.00 © 2003 IEEE


380 2003 PROCEEDINGS Annual RELIABILITY AND MAINTAINABILITY Symposium
2.0. SCALES OF MEASUREMENT which item is first, which second, and so on.
Ordinal measurement scales are frequently used. For
The essence of measurement is that it allows us to separate
example, ratings on a scale of 1 to 10 for such things as
things into categories based on the property we are
beauty, friendliness, and pain, are common applications.
measuring—as Lord Kelvin says, to “know something about
Another is grades for student work where an “A” paper is
it”. But all measurements do not allow for the same level of
better than a “B” paper but there is no implication that the
quantification; what we “know” and how we can use it
difference in quality between an “A” paper and a “B” paper is
depends on the scale of measurement. Four types of
the same as the difference between a “C” paper and a “D”
measurement scales are generally recognized [4].
paper.
• Nominal Measurements The operations of multiplication and division are not
• Ordinal Measurements meaningful on ordinal numbers and addition and subtraction
• Equal Interval Measurements while sometimes meaningful, must be carefully done since
• Ratio Measurements they assume an equal interval between the category labels.
2.1. Nominal Measurements For example, even though grades are ordinal numbers, the
Nominal measurements are the most basic type. The term grade-point-average is often used as an overall measure of a
“nominal” refers to the names of things and nominal student’s performance.
measurements use the name of a property to put objects into 2.3. Equal Interval Measurements
categories. For example, the property of shape can be Equal interval measurements are ones in which the
represented by such names as “square”, “rectangle”, “circle”, categories are ordered by the amount of the property they have
“triangle”, etc. and objects can be categorized by shape in this and the intervals between the categories are the same
way. Properties, measured by nominal measurements, can everywhere. The measurement of temperature on a Fahrenheit
only be used to put objects into categories that are different scale is an example of an interval measurement. The
from each other; questions of “more” or “less” have no difference between a temperature of 10º F and one of 11º F is
meaning. the same as the difference between 95º F and 96º F.
Numbers can be used as the category names, but such Addition, subtraction, and operations such as the calculation
numbers are only labels; they do not represent amounts and of an average are all meaningful operations for the numbers on
while it is possible to do arithmetical calculations on such an interval scale since the differences between categories are
numbers, the results are meaningless. For example, one the same everywhere. However, the placement of the zero
attribute of a student in a class is his or her the social security point on an interval scale is arbitrary and this means that the
number. These numbers can be entered into a spreadsheet and numbers cannot be meaningfully multiplied or divided. For
their average can be calculated, but this “average social example, the average of several temperature measurements
security number” is meaningless. On the other hand, the can be easily computed but it is meaningless to say that 60º F
average height or weight of the students in the class is is twice as warm as 30º F. This becomes very apparent if
meaningful. these temperatures are converted to Celsius (approximately
Nominal measurements are important in reliability because, 15.6º C and –1.1º C respectively). The Fahrenheit and Celsius
although we get no quantitative information from the scales differ not only in the size of the degree (the distance
measurement, the collection of such measurements provides between the category labels), but also in the location of the
useful information. For example, by counting the number of zero point. Because the zero point is arbitrary, measurements
observations of a piece of equipment that fall into the on interval scales often have negative values.
categories “working” and “failed” we can determine the
frequencies and the distributions of those states and in some 2.4. Ratio Measurements
cases what caused the equipment to change from one state to A ratio scale is the only scale on which the numbers can be
the other. meaningfully multiplied and divided. Like an interval scale,
the intervals between categories are equal, but a ratio scale
2.2. Ordinal Measurements also has a true zero-point indicating that absolutely none of the
The term “ordinal” refers to the order of things. Ordinal property being measured is present. It is not possible to have
measurements specify the order in which things are to be less than zero on a ratio scale. Since it has a true zero, it is
placed from first to last, but they do not specify how far apart reasonable to multiply and divide the numbers on the scale.
the things are. Ordinal measurements tell us how much of a The Kelvin scale for measuring temperature is a ratio scale.
property an item has, but only relative to another item. Thus, The intervals are the same as on the Celsius scale but 0º K
using ordinal measurements, we can say that one item has represents no heat whatsoever and it is not possible to have a
more or less of some property than another, but not how much temperature of less than 0º. A temperature of 1000º K is twice
more or how much less. as hot as a temperature of 500º K.
Numbers, and sometimes letters of the alphabet or words, In reliability work, the time that a device has been
are often used as the symbols for ordinal measurements. Such operational is an example of a ratio measurement. A device
symbols have a natural order, which can be associated with the that fails after 10,000 hrs of operation has lasted twice as long
amount of the property they represent. These category as one that failed after 5,000 hrs.
numbers are not arbitrary labels as in the case of nominal
measurements; instead they serve to identify ranks, specifying 2.5. Linear and Logarithmic Scales
Interval and ratio measurement results are most often

2003 PROCEEDINGS Annual RELIABILITY AND MAINTAINABILITY Symposium 381


reported on either a linear or a logarithmic scale. Table 1. Severity Evaluation Criteria [3, p 11]
Linear scales are most appropriately used when adding 1 Rank Effect Criteria: Severity of Effect
unit to a value represents as big a change in the quantity being 10 Hazardous Very high severity ranking when a potential failure mode
without affects safe vehicle operation and/or involves
measured regardless of its original size. Thus, on a linear warning noncompliance with government regulations without
scale a change of 1 is the same regardless of whether the warning.
original value was 1 or 1000. Linear scales often cover a 9 Hazardous Very high severity ranking when a potential failure mode
limited range of values—partly because the “equivalence of a with affects safe vehicle operation and/or involves
unit change anywhere” property may only hold over a limited warning noncompliance with government regulation with
warning.
range of values and partly because the size of the scale’s 8 Very High Vehicle/item inoperable, with loss of primary function.
physical representation may restrict its range. 7 High Vehicle/item operable, but at reduced level of
If doubling a value is just as big a change, whether it is from performance. Customer dissatisfied.
1 to 2 or 1000 to 2000, a logarithmic scale is appropriate. On 6 Moderate Vehicle/item operable, but comfort/convenience item(s)
inoperable. Customer experiences discomfort.
a logarithmic scale proportional changes in a value appear as
5 Low Vehicle/item operable, but comfort/convenience item(s)
the same difference. Logarithmic scales also are able to operable at reduced level of performance. Customer
compactly represent a very large range of values. Many experiences some dissatisfaction.
natural things have these characteristics and logarithmic scales 4 Very Low Cosmetic defect in finish, fit & finish/squeak or rattle
are used for their measurement. For example, the pH measure item that does not conform to specifications. Defect
noticed by most customers.
of the acidity of a liquid is the negative of the logarithm of the
3 Minor Cosmetic defect in finish, fit & finish/squeak or rattle
concentration of hydrogen ions. An acid with a pH of 3 has a item that does not conform to specifications. Defect
concentration of 0.001 (10-3) ions, neutral water has a pH of 7 noticed by average customer.
(10-7) and a strong base might have a pH of 11 (10-11). Such a 2 Very Cosmetic defect in finish, fit & finish/squeak or rattle
range of values can hardly be plotted on a linear scale. As Minor item that does not conform to specifications. Defect
noticed by discriminating customer.
another example, graphs of the prices of stocks are usually put 1 None No effect.
on a logarithmic scale since a doubling in the stock’s value per
share from $1 to $2 is as important as a doubling from $20 to might be debated and whether a vehicle running so poorly that
$40. it is capable of going only 10 miles per hour is “inoperable” or
In reliability work the probability of failure is often “operable at a reduced level of performance” might also be a
represented on a logarithmic scale. Failure probabilities often matter of individual judgment.
cover a range of 10 or more orders of magnitude from less The numerical ranks serve to order the severity assessments
than 10-9 to 1. Also, an order of magnitude difference in from low to high as on an ordinal scale. Things that affect
reliability is usually needed to have a meaningful difference. only the appearance of the vehicle are ranked lower than those
3. COMPONENTS OF THE RPN that affect its operation and those, in turn, are ranked lower
than things that affect the safety of the vehicle operation.
We now examine the components of the RPN calculation, While this ordering seems subjective, most people would
severity, occurrence, and detection, with respect to the types probably agree with it: safety is more important than
of scales on which they are measured. operation, and operation is more important than appearance,
3.1. Severity the severity scale does not appear to have any underlying scale
Severity is a measurement of the seriousness of the effect of of measurement. If cost were the underlying criteria, some
the potential failure mode on the next higher assembly, the cosmetic defects, such as a molted appearance that required a
system, or the customer. The SAE J1739 procedure vehicle to be repainted would be ranked substantially higher
recommends that the effects of the failure mode occurrence be than some operational defects, such as noisy brakes that could
described as the effect on the customer or as they would be be fixed by using a different type of brake pad.
noticed or experienced by the customer. Common failure Clearly, the scale is not an interval scale. The difference in
effects are such things as: whether a user is warned of a hazardous condition or not
• Excessive noise seems to be (in some sense) more important than the
• Erratic or intermittent operation difference in whether an “average” or a “discerning” customer
• Instability would notice a cosmetic defect.
• Impaired control Thus, the severity scale is at most an ordinal scale with the
• Unpleasant odor ordering based on a subjective judgment of “severity” and
• Poor appearance designated by the assigned ranks (1 to 10).

Table 1 shows the criteria used to rank the severity of a 3.2. Occurrence
failure effect. Occurrence is an assessment of the likelihood that a specific
The category descriptions for the severity scale are detailed cause or mechanism of a failure mode occurs. Some typical
enough that with minimal training and a good set of examples failure causes and failure mechanisms are:
most people would be able to agree on which category a given • Incorrect material specification
failure should be classified, or at least they would put them in • Over-stressing
nearby categories. Whether something would be noticed by a • Insufficient lubrication capability
“discerning customer” but not by an “average customer” • Inadequate maintenance instructions

382 2003 PROCEEDINGS Annual RELIABILITY AND MAINTAINABILITY Symposium


• Fatigue 10 10
• Material instability 8
Table 2 lists the occurrence rankings and their associated 6
meanings. The “Design Life Possible Failure Rates” is the R
Rank
r
relative number of failures anticipated during the design life of 4
the item. Estimates of the probability of occurrence may be
2
based on the service history and field experience of similar
components with appropriate changes for differences in the 0 0
-77 F 10-3
design or operating environment. 10
10 10-5 r 0.1 11
Probability of Failure
Table 2. Occurrence Ranking Evaluation Criteria [3, p.15]. Figure 1. Rank vs. Probability of Failure for the Occurrence
Ranking Probability of Failure Design Life Possible Failure Rates scale.
10 Very High: Failure is almost ≥0.5, (greater than 1 in 2)
inevitable
9 0.33, (1 in 3) 3.3. Detection
8 High: Repeated failures 0.125, (1 in 8) Detection assumes that a company has established some
7 0.05, (1 in 20) kind of “design verification program” that is able to determine
6 Moderate: Occasional 0.0125 (1 in 80)
Failures
whether or not an identified potential cause of failure will
5 0.0025 (1 in 400) actually result in the indicated failure mode. The design
4 0.0005 (1 in 2000) controls used in such a program include those that:
3 Low: Relatively few failures 0.0000667 (1 in 15,000) • Prevent the failure cause or mechanism or the failure
2 0.00000667, (1 in 150,000) mode or its effect from occurring, or reduce their rate of
1 Remote: Failure is unlikely ≤0.000000667, (1 in 1,500,000) occurrence.
The occurrence rankings clearly have an underlying • Detect the failure cause or mechanism and lead to
measurement of the probability of the failure mode occurring. corrective actions, and
This can be estimated statistically from data or from a variety • Detect the failure mode.
of system models. Figure 1 shows the rank versus the Detection is an assessment of the ability of such a program
probability of failure from Table 2 plotted on a log scale*. If to identify a potential weakness before the part or assembly is
the rank represented the log of the probability of failure, the released to production. Examples of the types of tests used for
steps in Figure 1 would all be of uniform width. Since they this purpose in the automobile industry include: road tests,
are not, it is easily observed that a change of 1 rank at the high design reviews, simulations, mathematical studies, laboratory
end of the scale (e.g., from 7 to 8) represents a proportionally tests, feasibility reviews, prototype tests, and fleet tests. Table
a smaller change in the probability of occurrence than a 3 shows the evaluation criteria used for the detection rankings.
change in rank at the low end of the scale (e.g., from 2 to 3). Table 3. Detection Ranking Evaluation Criteria [3, p. 17]
Thus, the high end of the scale has a finer granularity than the Rank Detection Criteria: Likelihood of Detection by Design Control
low end of the scale. 10 Absolute Design Control will not and/or cannot detect a potential
Placing a straight edge along the edge of the steps as shown Uncertainty cause/mechanism and subsequent failure mode; or there
in Figure 1, shows that below rank 6 the ranks are roughly is no Design Control.
9 Very Very remote chance the Design Control will detect a
proportional to the log of the probability of failure. Thus, the Remote potential cause/mechanism and subsequent failure mode.
occurrence scale provides an interval measure for the lower 8 Remote Remote chance the Design Control will detect a
ranks where a change in rank represents a constant potential cause/mechanism and subsequent failure mode.
proportional change in the underlying probability of failure. 7 Very Low Very low chance the Design Control will detect a
For the higher ranks this is not true and the scale is an ordinal potential cause/mechanism and subsequent failure mode.
6 Low Low chance the Design Control will detect a potential
scale in which higher numbered ranks indicate a higher cause/mechanism and subsequent failure mode.
probability of failure but the relative increase signified by a 5 Moderate Moderate chance the Design Control will detect a
unit change in rank varies. potential cause/mechanism and subsequent failure mode.
4 Moderately Moderately high chance the Design Control will detect a
High potential cause/mechanism and subsequent failure mode.
3 High High chance the Design Control will detect a potential
cause/mechanism and subsequent failure mode.
* 2 Very High Very high chance the Design Control will detect a
Figure 1 interprets Table 2 such that the given probability of occurrence for
potential cause/mechanism and subsequent failure mode.
each rank is the highest permissible for that rank, i.e., rank 1 failures have a
1 Almost Design Control will almost certainly detect a potential
highest probability of occurrence of 6.67x 10-7 (1/1,500,000), rank 2 failures
Certain cause/mechanism and subsequent failure mode.
from 6.67x10-7 to 6.67 x 10-6, etc. This interpretation leaves a discrepancy at
rank 10, where rank 10 failures are plotted as having a probability of The detection ranking is clearly a subjective judgment on the
occurrence of greater than 1/3 instead of greater than 1/2 as shown in the
table. If we had started with rank 10 and assumed the given probabilities were part of the designer as to how likely it is that a failure mode
the lower bound for each level there would have been a discrepancy at rank 1. will be found in subsequent tests. Although consistent
Still other interpretations are possible for example the values in Table 2 could estimates might be possible with training one would expect
be interpreted as midpoints of the range for each rank. Nevertheless, the that there would be a great deal of variation particularly
particular interpretation used does not materially affect the discussion or the
conclusions reached. among the middle ranks. Thus, the detection scale is an

2003 PROCEEDINGS Annual RELIABILITY AND MAINTAINABILITY Symposium 383


ordinal scale, ordered by the “likelihood” of the design 24 different combinations of S, O, and D. Similarly, seven
verification program detecting the weakness. One would RPN numbers: 24, 36, 40, 48, 80, 90, and 180, can be formed
expect that similar situations might be rated quite differently from 21 different combinations. Figure 2 shows the number
by different people or design teams. of ways each number on the RPN scale can be formed.
It is difficult to accept that failures whose severity ranges
4.0. THE RPN
Severity Occurrence Detection
As already noted the RPN is calculated as the product of the
1 8 8
rankings assigned to each factor: Severity (S), Occurrence (O),
2 4 8
and Detection (D), and the result lies on an integer scale from
2 8 4
1 to 1000.
4 2 8
4.1. Holes in the Scale 4 4 4
But the RPN scale is not continuous, indeed it has many 4 8 2
“holes”—that is, many of the numbers in the range of 1 to 8 1 8
1000 cannot be formed from the product of S, O, and D. This 8 2 4
is, perhaps, the most serious of the problems in interpreting 8 4 2
the RPN. While it is true that the numbers cover a range from 8 8 1
1 to 1000, 88% of that range is empty; only 120 of the 1000 Figure 3. Combinations of S, O, and D yielding an RPN of 64.
numbers generated from the product of S, O, and D are
unique. No number having a prime factor greater than 10 can from 1 (not noticeable except by the most discerning
be formed from the product of S, O, and D. Thus, the customer) to 8 (inoperable with loss of primary function) can
numbers 11, 22, 33, ... 990, which are all multiples of 11 be evaluated as having the same importance. This is even
cannot be formed and are excluded. Similarly, all multiples of more so for RPNs like 60 for which the severity factor can
13, 17, 19, etc. are excluded. 1000 is the largest number, but vary from 1 to 10 (hazardous without warning). If nothing
900 is the second largest followed by 810, 800, 729, and 720. else, the need for assurance that a failure rate is truly
Figure 2 shows the numbers formed by the RPN and the acceptably low is much greater when the severity of the failure
“holes” in the scale between the numbers graphically. effect is higher.
To pursue this observation a little further, 4 is near the We note that 64 is the product 4 x 4 x 4; if we increase only
middle of each scale; the next number after 64 (4 x 4 x 4) is one factor to 5 (i.e., 5 x 4 x 4) we get an RPN of 80. The
70. None of the numbers 65, 66, 67, 68, and 69 can be formed numbers between 64 and 80 can be formed in a variety of
from any combination of S, O, and D. In the RPN analysis, ways. One of these, 70 can result from the combination S =
high numbers are “more important” than low numbers, but the 10, O = 7, D = 1. Again, it is difficult to accept that a failure
presence of holes makes it difficult to interpret what the mode having a severity of 10 (hazardous without warning), a
differences between the numbers mean. Since 70 is the next high probability of occurrence (7), and a low detection (1) can
number after 64, a valid question is: “Is the difference between be bounded in importance by systems having moderate (4 or
63 and 64 the same or less than the difference between 64 and 5) severity, moderate (4) occurrence, and moderate (4)
70?” How about the difference between 900 and 1000. detection.
Allowing an analyst to assign values between the ranks using
small fractions such as ½ or ¼ on the ranking scales (e.g., a 4.3. Sensitivity to Small Changes
rank of 5½) would make the holes smaller but would not solve Multiplying the numbers comprising the RPN is intended to
the problem. magnify the effects of high risk factors. This also makes the
25
RPN such that a small change in one factor has a much larger
effect when the other factors are larger than when they are
15
small. Hence, small variations in one ranking can lead to very
different effects on the RPN, depending on the values of the
5 other factors. If D and O are both 10, a 1 point difference in
0 the severity ranking results in a 100 point difference in the
11 101 220
0 01 3 01 4
400 01 501 6
600 01 7 01 8 0 01
80 90 1 1000
RPN; at the other extreme of D and O equal to 1, the same 1
Figure 2. RPN scale showing the number of occurrences of point difference only gives a 1 point difference in the RPN;
each number. near the middle of the scales, if D and O are both 4, a 1 point
difference gives a 16 point difference in the RPN. We also
4.2. Duplicate RPN Numbers note that the 100 point swing due to a change in S (with O and
Since 1000 numbers are produced from the product of S, O, D both 10) could occur anywhere on the RPN scale from a
and D but only 120 of them are unique, there must be many low of 200 (goes to 100) to a high of 1000 (goes to 900). As
duplicate numbers. In general each of the rankings can be noted earlier individual judgments may give small differences
formed in several different ways. (Only 6 RPNs are formed in the severity ranking and may result in larger differences in
by a single, unique combination of S, O, and D.) For example, the more subjective detection ranking. These small
as shown in Figure 3, 64 can be formed from 10 different differences in judgment can lead to potentially large
combinations of S, O, and D: differences in the RPN ranking for very similar failure modes.
The RPN numbers, 60, 72, and 120 can each be formed from Both of the combinations of S, O, and D below result in an

384 2003 PROCEEDINGS Annual RELIABILITY AND MAINTAINABILITY Symposium


RPN of 192. depending on the values of the other numbers. Thus, the RPN
number itself offers no clue as to how the design can be
Severity (S) Occurrence (O) Detection (D) RPN
improved or even how big a change in the number constitutes
3 8 8 192
a meaningful improvement. Comparisons of different RPNs
8 3 8 192
have little meaning and design improvements made to reduce
But a 1 point change in S in the first causes a 64 point change the RPN may, in fact, not be improvements at all.
in the RPN, whereas in the second a 1 point change in S The RPN approach should be dropped and an entirely
causes only a 24 point change in the RPN. different prioritization techniques should be used. The
following recommendations are made:
4.4. Comparing RPNs 1. The detection ranking should be dropped. (Others have
The RPN is an attempt to quantify and rank assessments of also recommended using only the severity and occurrence to
the suitability of a design on three dimensions: the severity of
prioritize the failure modes; see for example [5, 6, 7].) First,
the effect of a failure, the frequency of occurrence of the
the category name is misleading since it is a measure of
failure, and the likelihood of the failure mode detection, along
whether subsequent testing will show that the failure mode
a single dimension so that design concerns can be prioritized
actually exists rather than whether the failure will be detected
and compared. This is generally not possible without some when it occurs. Second, the rank assigned for the failure mode
kind of cost function that quantifies how reductions along one is highly subjective and one would expect considerable
dimension relate to changes along another dimension.
variation from one analyst or design team to another. (This
Calculation of the RPN implies that trade-offs can be made
could be offset with extensive training but it remains a
between the severity, occurrence, and detection factors
problem.) Third, a low detection ranking means that the
according to a geometric cost function. For example, doubling
company’s “design verification program” will be able to
the severity (e.g., 4 → 8), halving the occurrence, (e.g., 4 → 2) detect that the potential problem is real, but fixing design
and keeping the detection the same has no net effect on the flaws late in the design process is almost always expensive.
RPN. But for this property (e.g., doubling or halving the Even if a design flaw is “certain” to be found by a fleet test, or
values) to be meaningful the underlying scale of measurement a field test after the design is complete, that is much too late to
must be a ratio scale. The operations have no meaning on an be of much benefit. The detection ranking should be replaced
ordinal scale. with a comment indicating what tests might be needed to more
About the best that can be done to compare RPNs is a accurately determine the frequency of occurrence of the
multidimensional Pareto ranking in which a failure whose S, potential failure.
O, and D values are all lower than another’s is ranked lower 2. The rankings for the severity categories should also be
than the other. (E.g., RPN1 < RPN2 if S1 < S2, O1 < O2, and D1 scrapped. Instead these categories should be considered as
< D2). nominal classifications (categorizations) for safety,
The use of a single number such as the RPN to “Rank” operational, and cosmetic related failure mode effects. An
failure modes encourages management to set an arbitrary acceptable upper limit on the probability of occurrence should
lower threshold, e.g., "We will only worry about concerns be determined for each class and all failure modes in the class
having an RPN greater than 64," as a way of deciding what whose probability of occurrence was too high (above that
concerns will be accepted and which ones will be corrected. limit) would then need to be corrected or mitigated as
This is especially true for a large complex product, such as an appropriate. If the analysis is being done for safety the
automobile in which thousands of failure modes must be “Hazardous without warning” (Rank 10) and “Hazardous with
prioritized so that limited resources can be productively used warning” (Rank 9) and perhaps the “Very High” (Rank 8)
to correct potential problems. Unfortunately, the practical categories would be of the most interest. If the analysis is
effect of this thresholding can be an unproductive “numbers being done with respect to the operational performance and
games” in which the objective becomes to “get the RPN below reliability of the vehicle the “Very High” (Rank 8), “High”
the threshold” rather than to improve the quality of the design. (Rank 7), “Moderate” (Rank 6), and “Low” (Rank 5) would be
of the most interest. If the analysis is being done with respect
5.0. CONCLUSIONS AND RECOMMENDATIONS to the marketability of the automobile the cosmetic defect
Although the RPN methodology is very widely used it is a categories: “Very Low” (Rank 4), “Minor” (Rank 3), and
severely flawed ranking technique. The fundamental problem “Very Minor” (Rank 2) categories would also become of more
is that the scales used to order the failure modes in terms of concern. The overall design would have to be concerned with
the severity, occurrence, and detection ratings are ordinal all of these types of potential failures. It might be appropriate
scales but they are treated as if numerical operations on them, to rename these categories to better describe the types of
most notably multiplication, are meaningful. The results are failures they describe rather than to give them an implied
not only meaningless but are in fact misleading. ordinal ranking, which might not be applicable. A frequently
The RPN number itself is not on a continuous scale as might occurring “Minor” defect in a vehicle’s finish could make it
at first be perceived, but one with many holes and the same difficult to sell and give the manufacturer a reputation for
number can (in general) be constructed from many different “poor quality”.
combinations of the severity, occurrence, and detection If a cost could be associated with each failure effect, that
rankings. Small changes in one of the factors from which the would allow the failures to be placed on a dollar scale (a ratio
RPN is computed can have different effects on the RPN scale); the product of the cost of the failure effect and the

2003 PROCEEDINGS Annual RELIABILITY AND MAINTAINABILITY Symposium 385


probability of occurrence of the underlying failure mode could 7. P. Palady, M. Horvath, and C. Thomas, “Restoring the Effectiveness of
then be multiplied to give an “expected cost” of the failure. Failure Modes and Effects Analysis”, Society of Automotive Engineers
Proposed design changes could then be evaluated by their International Congress and Exposition, Detroit, MI, 1994, Technical Paper
No. 940884.
effect on the expected cost.
8. P. J. Fleming and J. J. Wallace, “How not to lie with statistics: the correct
way to summarize benchmark results”, Communications of the ACM, Vol.
REFERENCES 29, No. 3, March 1986, pp. 218-221.
1. http://www.top-biography.com/9103-
William%20Thomson/quotations.htm.
2. "Potential Failure Mode and Effects Analysis In Design (Design FMEA) BIOGRAPHY
and For Manufacturing and Assembly Processes (Process FMEA) John B. Bowles
Instruction Manual", Ford Motor Company, September 1988. Computer Science and Engineering
University of South Carolina
3. "Potential Failure Mode and Effects Analysis In Design (Design FMEA) Columbia, SC 29208 USA
and Potential Failure Mode and Effects Analysis In Manufacturing and e-mail: bowles@engr.sc.edu
Assembly Processes (Process FMEA) Reference Manual", Society of
Automotive Engineers, Surface Vehicle Recommended Practice, J1739, John Bowles is an Associate Professor in the Computer Science and
July 1994. Engineering Department at the University of South Carolina where he teaches
4. C. C. Reaves, Quantitative Research for the Behavioral Sciences, John and does research in reliable system design. Previously he was employed by
Wiley, New York, NY, 1992. NCR Corporation and Bell Laboratories where he worked on several large
system design projects. He has a BS in Engineering Science from the
5. "Recommended Failure Mode and Effects Analysis (FMEA) practices for University of Virginia, an MS in Applied Mathematics from the University of
non-automobile applications", Society of Automotive Engineers, Michigan, and a Ph.D. in Computer Science from Rutgers University.
Aerospace Recommended Practice, ARP 5580, July 2001.
6. J. B. Bowles, "The new SAE FMECA standard", Proc. Ann. Reliability and
Maintainability Symp., Anaheim, CA, January 1998, pp. 48-53.

386 2003 PROCEEDINGS Annual RELIABILITY AND MAINTAINABILITY Symposium

S-ar putea să vă placă și