Availability (system)

Availability is the probability that a system will work as required when required during the period of a mission. The mission could be the 18-hour span of an aircraft flight. The mission period could also be the 3 to 15 month span of a military deployment. Availability includes non-operational periods associated with reliability, maintenance, and logistics.

This is measured in terms of nines. Five-9's means less than 5 minutes when the system is not operating correctly over the span of one year.

Availability is only meaningful for supportable systems. As an example, availability of 99.9% means nothing after the only known source stops manufacturing a critical replacement part.


There are two kinds of availability.

  • Operational
  • Predicted

Operational availability is presumed to be the same as predicted availability until after operational metrics become available.


Operational availability is based on observations after at least one system has been built. This usually begins with the brassboard system that is used to complete system development, and continues with the first of kind used for live fire test and evaluation (LFTE). Organizations responsible for maintenance use this to evaluate the effectiveness of the maintenance philosophy.

Predicted availability is based on a model of the system before it is built.

Downtime is the total of all of the different contributions that compromise operation. For modeling, these are different aspects of the model, such as human-system interface for MTTR and reliability modeling for MTBF. For observation, these reflect the different areas of the organization, such as maintenance personnel and documentation for MTTR, and manufacturers and shippers for MLDT.


Mean Time Between Failure (MTBF) depends upon the maintenance philosophy.

If a system is designed with both redundancy and automatic fault bypass, then MTBF is the anticipated lifespan of the system if these features cover all possible failure modes (infinity for all practical purposes). Such systems will continue without noticeable interruption when these conditions are satisfied unless there are secondary failures. This is called active redundancy, which requires no maintenance to prevent mission failure. Active redundancy is required for systems that cannot be maintained, such as satellites.

If a system has no redundancy, then MTBF is the inverse of failure rate, .

Systems with spare parts that are energized but that lack automatic fault bypass are not actually redundant because human action is required to restore operation after every failure. This depends upon Condition-based maintenance and Planned Maintenance System support.


Mean Time To Recover (MTTR) is the length of time required to restore operation to specification.

This includes three values.

  • Mean Time To Discover
  • Mean Time To Isolate
  • Mean Time To Repair

Mean Time To Discover is the length of time that transpires between when a failure occurs and the system users become aware of the failure. There are two maintenance philosophies associated with Mean Time To Discover.

CBM works like your car where an oil indicator tells you when oil pressure is too low and a temperature indicator tells you when engine temperature is too high. There is zero time to discover a failure where an indicator is placed in front of a system operator.

PMS is required for silent failures that lack CBM. PMS works is periodic maintenance, like when you perform diagnostic tests on your car every 90 days (or 3,000 miles). A failure may occur any time during the 90 days, such as a broken light, but you will not become aware until you perform diagnostic test.

Mean Time To Discover is statistical when PMS is the dominant maintenance philosophy. For example, if a fault is discovered during PMS diagnostic procedure that is run every 10 days, the average fault duration will be 5 days. This creates a dependency between availability performance and labor costs. There is no such dependency associated with CBM.

Mean Time To Isolate is the average length of time required to identify a setting that needs to be adjusted or a component that needs to be replaced. This is dependent on documentation, training, and technical support. This tends to be less on systems that have CBM because users can begin with the list of items connected to the indicator used to notify users about the fault. This also tends to be less on fully documented systems.

Mean Time To Repair is the average length of time to restore operation. For mission critical systems, this is generally estimated by dividing time required to replace all parts by the number of replaceable parts.


Mean Logistics Delay Time is the average time required to obtain replacement parts from the manufacturer and transport those parts to the work site.


Mean Active Maintenance Down Time is associated with Planned Maintenance System support philosophy. This is average amount of time while the system is not 100% operational because of diagnostic testing that requires down time.

For example, an automobile that requires 1 day of maintenance every 90 days has a Mean Active Maintenance Down Time of just over 1%.

This is separate from the type of down time associated with repair activities.


Availability expectations are described in terms of nines.

The following table shows the anticipated down-time for different availabilities for a mission time of one year. This is the typical time-span used with commercial systems.

90% 99% 99.9% 99.99% 99.999% 99.9999%
40 days 4 days 9 hours 50 minutes 5 minutes 30 seconds


Systems that require maintenance are said to be supportable if they satisfy the following criteria.

  • Adequate labor for PMS diagnostics or fault indicators for failure modes that lack PMS
  • Access to manufacturing sources for replacement parts
  • Sufficient resources to sustain access to replacement parts and maintenance labor (money, lodging, ...)
  • Training and maintenance documentation suitable to restore operation
  • Transportation services to deliver replacement parts

Systems that lack any of these requirements are said to be unsupportable.

Mission Failure

Mission failure is the result of trying to use a system in its normal mode when it is not working.

Apart from human error, mission failure results from the following causes.

  • Lack of instrumentation for the fault in the form of an indicator suitable for CBM
  • Failure after PMS diagnostic testing has been completed but before operation is attempted
  • Lack of diagnostic PMS testing for any failure mode that lacks CBM

See also


 This article incorporates public domain material from the General Services Administration document "Federal Standard 1037C" (in support of MIL-STD-188).

External links

  This article uses material from the Wikipedia page available here. It is released under the Creative Commons Attribution-Share-Alike License 3.0.


Connect with defaultLogic
What We've Done
Led Digital Marketing Efforts of Top 500 e-Retailers.
Worked with Top Brands at Leading Agencies.
Successfully Managed Over $50 million in Digital Ad Spend.
Developed Strategies and Processes that Enabled Brands to Grow During an Economic Downturn.
Taught Advanced Internet Marketing Strategies at the graduate level.

Manage research, learning and skills at defaultLogic. Create an account using LinkedIn or facebook to manage and organize your IT knowledge. defaultLogic works like a shopping cart for information -- helping you to save, discuss and share.

  Contact Us