Académique Documents
Professionnel Documents
Culture Documents
Contents
Click on a section to jump to it
Introduction
Conclusion
Resources
10
Introduction
Mean time between failure (MTBF) has been used for over 60 years as a basis for various
decisions. Over the years more than 20 methods and procedures for lifecycle predictions
have been developed. Therefore, it is no wonder that MTBF has been the daunting subject of
endless debate. One area in particular where this is evident is in the design of mission
critical facilities that house IT and telecommunications equipment. When minutes of downtime can negatively impact the market value of a business, it is crucial that the physical
infrastructure supporting this networking environment be reliable. The business reliability
target may not be achieved without a solid understanding of MTBF. This paper explains
every aspect of MTBF using examples throughout in an effort to simplify complexity and
clarify misconception.
What is a failure?
What are the
assumptions?
These questions should be asked immediately upon reviewing any MTBF value. Without the
answers to these questions, the discussion holds little value. MTBF is often quoted without
providing a definition of failure. This practice is not only misleading but completely useless.
A similar practice would be to advertise the fuel efficiency of an automobile as miles per
tank without defining the capacity of the tank in liters or gallons. To address this ambiguity,
one could argue there are two basic definitions of a failure:
1. The termination of the ability of the product as a whole to perform its required function. 1
2. The termination of the ability of any individual component to perform its required function but not the termination of the ability of the product as a whole to perform.
The following two examples illustrate how a particular failure mode in a product may or may
not be classified as a failure, depending on the definition chosen.
Example 1:
If a redundant disk in a RAID array fails, the failure does not prevent the RAID array from
performing its required function of supplying critical data at any time. However, the disk
failure does prevent a component of the disk array from performing its required function of
supplying storage capacity. Therefore, according to definition 1, this is not a failure, but
according to definition 2, it is a failure.
Example 2:
If the inverter of a UPS fails and the UPS switches to static bypass, the failure does not
prevent the UPS from performing its required function which is supplying power to the critical
load. However, the inverter failure does prevent a component of the UPS from performing its
required function of supplying conditioned power. Similar to the previous example, this is
only a failure by the second definition.
If there existed only two definitions, then defining a failure would seem rather simple.
Unfortunately, when product reputation is on the line, the matter becomes almost as complicated as MTBF itself. In reality there are more then two definitions of failure, in fact they are
infinite. Depending on the type of product, manufacturers may have numerous definitions of
failure. Manufacturers that are quality driven track all modes of failure for the purpose of
process control which, among other benefits, drives out product defects. Therefore, additional questions are needed to accurately define a failure.
IEC-50
IEC-50
White Paper 78
Rev 1
Reliability,
availability,
MTBF, and MTTR
defined
MTBF impacts both reliability and availability. Before MTBF methods can be explained, it is
important to have a solid foundation of these concepts. The difference between reliability and
availability is often unknown or misunderstood. High availability and high reliability often go
hand in hand, but they are not interchangeable terms.
Reliability is the ability of a system or component to perform its required functions under
stated conditions for a specified period of time [IEEE 90].
In other words, it is the likelihood that the system or component will succeed within its
identified mission time, with no failures. An aircraft mission is the perfect example to illustrate
this concept. When an aircraft takes off for its mission, there is one goal in mind: complete
the flight, as intended, safely (with no catastrophic failures).
Availability, on the other hand, is the degree to which a system or component is operational
and accessible when required for use [IEEE 90].
It can be viewed as the likelihood that the system or component is in a state to perform its
required function under given conditions at a given instant in time. Availability is determined
by a systems reliability, as well as its recovery time when a failure does occur. When
systems have long continuous operating times (for example, a 10-year data center), failures
are inevitable. Availability is often looked at because, when a failure does occur, the critical
variable now becomes how quickly the system can be recovered. In the data center example, having a reliable system design is the most critical variable, but when a failure occurs,
the most important consideration must be getting the IT equipment and business processes
up and running as fast as possible to keep downtime to a minimum.
MTBF, or mean time between failure, is a basic measure of a systems reliability. It is
typically represented in units of hours. The higher the MTBF number is, the higher the
reliability of the product. Equation 1 illustrates this relationship.
Equation 1:
Reliability = e
Time
MTBF
White Paper 78
Rev 1
White Paper 78
Rev 1
Availability =
MTBF
( MTBF + MTTR)
For Equation 1 and Equation 2 above to be valid, a basic assumption must be made when
analyzing the MTBF of a system. Unlike mechanical systems, most electronic systems dont
have moving parts. As a result, it is generally accepted that electronic systems or components exhibit constant failure rates during the useful operating life. Figure 1, referred to as
the failure rate bathtub curve, illustrates the origin of this constant failure rate assumption
mentioned previously. The "normal operating period" or useful life period" of this curve is the
stage at which a product is in use in the field. This is where product quality has leveled off to
a constant failure rate with respect to time. The sources of failures at this stage could include
undetectable defects, low design safety factors, higher random stress than expected, human
factors, and natural failures. Ample burn-in periods for components by the manufacturers,
proper maintenance, and proactive replacement of worn parts should prevent the type of
rapid decay curve shown in the "wear out period". The discussion above provides some
background on the concepts and differences of reliability and availability, allowing for the
proper interpretation of MTBF. The next section discusses the various MTBF prediction
methods.
Early failure
period
Figure 1
Bathtub curve to illustrate
constant rate of failures
Normal operating
period
Wear out
period
Rate
of
failure
Methods of
predicting and
estimating MTBF
Time
Oftentimes the terms prediction and estimation are used interchangeably, however this is
not correct. Methods that predict MTBF, calculate a value based only on a system design,
usually performed early in the product lifecycle. Prediction methods are useful when field
data is scarce or non-existent as is the case of the space shuttle or new product designs.
When sufficient field data exists, prediction methods should not be used. Rather, methods
that estimate MTBF should be used because they represent actual measurements of failures.
Methods that estimate MTBF, calculate a value based on an observed sample of similar
systems, usually performed after a large population has been deployed in the field. MTBF
estimation is by far the most widely used method of calculating MTBF, mainly because it is
based on real products that are experiencing actual usage in the field.
All of these methods are statistical in nature, which means they provide only an approximation of the actual MTBF. No one method is standardized across an industry. It is, therefore,
critical that the manufacturer understands and chooses the best method for the given
White Paper 78
Rev 1
Cushing, M., Krolewski, J., Stadterman, T., and Hum, B., U.S. Army Reliability Standardization
Improvement Policy and Its Impact, IEEE Transactions on Components, Packaging, and Manufacturing
Technology, Part A, Vol. 19, No. 2, pp. 277-278, 1996
White Paper 78
Rev 1
White Paper 78
Rev 1
Related resource
White Paper 78
Rev 1
Conclusion
MTBF is a buzz word commonly used in the IT industry. Numbers are thrown around
without an understanding of what they truly represent. While MTBF is an indication of
reliability, it does not represent the expected service life of the product. Ultimately an MTBF
value is meaningless if failure is undefined and assumptions are unrealistic or altogether
missing.
White Paper 78
Rev 1
Resources
Click on icon to link to resource
References
1. Pecht, M.G., Nash, F.R., Predicting the Reliability of Electronic Equipment, Proceedings of the IEEE, Vol. 82, No. 7, July 1994
2. Leonard, C., MIL-HDBK-217: Its Time To Rethink It, Electronic Design, October 24,
1991
3. http://www.markov-model.com
4. MIL-HDBK-338B, Electronic Reliability Design Handbook, October 1, 1998
5. IEEE 90 Institute of Electrical and Electronics Engineers, IEEE Standard Computer
Dictionary: A Compilation of IEEE Standard Computer Glossaries. New York, NY:
1990
Contact us
For feedback and comments about the content of this white paper:
Data Center Science Center, APC by Schneider Electric
DCSC@Schneider-Electric.com
If you are a customer and have questions specific to your data center project:
Contact your APC by Schneider Electric representative
White Paper 78
Rev 1
10