Skip to main content


Your product has little or no value if its functionality isn't available to users. Bugs, crashes, and network outages are examples of what might make your product's functionality unavailable at times.

Product managers therefore typically attach an availability constraint (nonfunctional requirement) to each functional requirement of the product. If one of the functions of the product is to generate reports, for example, a product manager should specify how likely it should be at any particular time that a user will be able to use this functionality.

The question with nonfunctional requirements is always the metric - how you measure them. How do you measure availability? Here are some options:
  • mean-time-between-failures (MTBF) - the average amount of time elapsed between failures to deliver the functionality.
  • failure rate - the frequency that the product fails to deliver the functionality. Failure rate is the reciprocal of MTBF, and often is expressed in terms of failures per hour.
  • uptime - percentage of the time that the functionality is available.
  • downtime - percentage of the time that the functionality is not available.
  • mean-time-between-system-abort (MTBSA) - the average amount of time elapsed between complete "reboots" of the system.
  • mean-time-between-critical-failure (MTBCF) - distinguishes between critical and noncritical failures.
  • maintenance free operating period (MFOP) - the average amount of time that the functionality is available without any special intervention or system maintenance.
Of course, a prospective customer will always want 100% uptime, but such availability is typically not practical to achieve. If you base a contract on 100% uptime, you will almost certainly be in violation of your contract at some point.

UPDATE: Scott Sehlhorst adds a number of important observations in this entry's comments. One thing he notes is that I neglected to mention MTTR:
  • mean-time-to-repair (MTTR) - the average amount of time it takes to repair the system after its functionality becomes unavailable. For hardware products, it usually refers to the time to replace a module or part. For software products, it can refer to the amount of time it takes to reboot or restart the system.
Also, some people use "availability" to refer strictly to uptime, and consider all of these parameters to be "reliability" metrics.


Roger L. Cauvin said…
Sounds like you have some good experience with the various metrics, Scott. I did leave out MTTR, I am going to update the entry to include it.

I also do think it's important to provide context to these metrics. You can use a product in many different circumstances; the availability requirements should specify the possible circumstances to the extent practicable.

As I've stated before, just finding - or even exploring - the right combination of metrics is arguably more important than assigning all of the exact numbers and exhaustive circumstances. (When there's a contract involved, these other factors are obviously still very important.)

Popular posts from this blog

Why Spreadsheets Suck for Prioritizing

The Goal As a company executive, you want confidence that your product team (which includes all the people, from all departments, responsible for product success) has a sound basis for deciding which items are on the product roadmap. You also want confidence the team is prioritizing the items in a smart way. What Should We Prioritize? The items the team prioritizes could be features, user stories, epics, market problems, themes, or experiments. Melissa Perri  makes an excellent case for a " problem roadmap ", and, in general, I recommend focusing on the latter types of items. However, the topic of what types of items you should prioritize - and in what situations - is interesting and important but beyond the scope of this blog entry. A Sad but Familiar Story If there is significant controversy about priorities, then almost inevitably, a product manager or other member of the team decides to put together The Spreadsheet. I've done it. Some of the mos

Interaction Design: the Neglected Skill

Your product development organization has a big, gaping hole in it. (Be prepared to feel defensive as you continue reading.) One of the most important roles in product development is the role of interaction designer. An interaction designer designs how the users will interact with the product and conceptualize the tasks they perform. He decides whether, for example, the user interface will be command driven, object oriented (clicking on objects then specifying what to do with them), or wizard based. The interaction designer decides the individual steps in the use cases. Every company has one or more people that play the interaction designer role. Usually, those people have little or no expertise in interaction design. Sadly, they typically don't even realize how unqualified they are. Let's see who typically plays the role at companies. Engineer . An engineer is an expert on building what is designed. Yes, an engineer may know how to design the internal structure of the hardware

Stop Validating and Start Falsifying

The product management and startup worlds are buzzing about the importance of "validation". In this entry, I'll explain how this idea originated and why it's leading organizations astray. Why Validate? In lean startup circles, you constantly hear about "validated learning" and "validating" product ideas: The assumption is that you have a great product idea and seek validation from customers before expending vast resources to build and bring it to market. Indeed, it makes sense to transcend conventional approaches to making product decisions . Intuition, sales anecdotes, feature requests from customers, backward industry thinking, and spreadsheets don't form the basis for sound product decisions. Incorporating lean startup concepts , and a more scientific approach to learning markets, is undoubtedly a sounder approach. Moreover, in larger organizations, sometimes further in the product life-cycle, everyone seems to have an opinio