Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
essential; after all, it takes a human being several seconds just<br />
to mentally process a problem.<br />
The point of this mental exercise was to convert times into more<br />
“tangible” measurements, which may be a surprise to those who<br />
haven’t thought about it before. If you found that frightening, the<br />
situation is even worse than presented above. Most of those<br />
problems presented above are the kind that the web host has<br />
some measure of control over. In real life, you have to worry<br />
about datacenters, communications lines, electricity, and a variety<br />
of other things that the host has little to no ability to manage.<br />
Even when conditions are ideal, there will always be the<br />
possibility of something catastrophic. All the preparation in the<br />
world can’t totally protect a single datacenter from something<br />
as rare and drastic as a hurricane, a terrorist attack, or a large<br />
meteorite hitting your servers. To the individual host, those<br />
events probably lend themselves to greater concerns than<br />
uptime, but that isn’t going to keep your clients from taking you<br />
to court over losses.<br />
So what’s the moral of the story? If you seriously want<br />
to approach 100% uptime, you’d better have redundancy,<br />
monitoring, and automation in place, with hosting infrastructure<br />
spread out across large areas of the country or planet.<br />
THE PROBABILITIES OF INDEPENDENT EVENTS<br />
Now that we’ve looked at some of the math and practical<br />
considerations of uptime percentages, most of what we’ve<br />
examined has been a bit of a downer. It’s time to turn the tables<br />
and make the math start working for the good guys, the hosts.<br />
Revisiting statistics, there is an important property in probabilities<br />
that is going to help more to achieve huge uptimes without being<br />
subject to the things that web hosts have no control over. That<br />
property is the statistical fact that the probability of independent<br />
events occurring is the product of their individual probabilities.<br />
To translate that into English, let’s say we have two servers<br />
with each having a 1% probability of being down. Assuming that<br />
downtime is independent, then we multiply the probabilities of<br />
each going down to get the odds of both being down. This works<br />
out to 1% times 1%, or 0.01%. To describe it another way, two<br />
servers may only be able to handle 99% uptime, but the odds<br />
that at least one of them are up is now 99.99% (assuming these<br />
events are independent).<br />
Incidentally, don’t try figuring this up by multiplying the uptimes.<br />
That leaves you with the probability of both being up and running,<br />
but we want to calculate the odds of at least one being up.<br />
Table 2 – Impacts of Redundancy on Uptime<br />
Uptime %<br />
90.0%<br />
99.0%<br />
99.9%<br />
1 Server<br />
90.0%<br />
99.0%<br />
99.9%<br />
2 Servers<br />
99.0000%<br />
99.9900%<br />
99.9999%<br />
3 Servers<br />
99.9000000%<br />
99.9999000%<br />
4 Servers<br />
99.9900000000%<br />
99.9999990000%<br />
99.9999999% 99.9999999999%<br />
If you examine the results of these probabilities (Table 2), there<br />
are some striking results. While a particular server may only<br />
provide 99.9% uptime, four independent servers working in<br />
tandem increase that uptime to 99.9999999999%. That’s going<br />
from three to twelve “nines,” which works out to going from 42<br />
minutes of downtime to mere milliseconds.<br />
The word that can’t be emphasized enough in this case is<br />
independent. That’s the only way this works.