StatusCake

Website Crashes Still a Major Issue in 2016

clock measuring uptime

Well, 2016 has been another banner year for major website crashes. Distributed denial of service (DDoS) attacks, such as the hugely disruptive one on Dyn, were on the rise, but many of these crashes were preventable. As in past years, the main culprit was inadequate preparation for an increase in traffic that was entirely predictable.

It is understandable that many new small businesses would not plan for traffic surges. However, just as in past years, large businesses and governmental agencies launched major campaigns that are guaranteed to generate huge increases in traffic and then found themselves the target of unrelenting negative publicity in the press and on social media when that increased traffic crashed their websites. As the late American baseball player Yogi Berra would say: “It’s déjà vu all over again.” Here are a just a few of the crashes that generated a lot of frustration.

Black Friday

Every major retailer knows that traffic on Black Friday will be off the charts as consumers chase the bargains that retailers had been previewing during the previous two weeks, so you would think they would be prepared. Unfortunately, some retailers were not. On Black Friday, the websites of the following companies crashed and/or experienced a severe degradation in service: Currys, PC World, Macy’s, Quidco (a cashback site), and GAME.

The negative reaction on social media was overwhelming. Frustrated customers vowed never again to shop at some of these retailers, and some helpfully posted links to competitors’ websites that were holding up just fine under the Black Friday onslaught.

Their poor performance on Black Friday hurt these firms in two ways. They lost an undetermined amount of sales to their competitors on the day because customers were unable to complete their purchases. Their reputation also suffered, causing them to lose future sales as well, Black Friday certainly was “black” for those firms, but for the wrong reasons.

Cabinet Office website crash

Leading up to the voting on the EU referendum in June, voters were encouraged to register online. Traffic on gov.uk steadily increased during the week before the registration deadline and traffic eventually reached over 200,000 users per hour. The site could not handle the traffic spike and crashed on the registration deadline date, leaving many people unable to become eligible to vote.

Frustrated voters took to social media to vent their anger and demand an extension of the voting registration deadline. After the furor, the government agreed to extend the registration deadline by two days.

As you can see, when a website goes down, users of the site can get very frustrated. Websites can go down for many reasons, not just because of a traffic spike. Sometimes technical issues at your hosting company over which you have no control can be the cause. In any event, you need to know promptly if your site is down so you can take corrective action and keep customer dissatisfaction to a minimum; consider using a website monitoring service to get alerted fast.

Share this

More from StatusCake

A Notification List Is Not a Team

3 min read In the previous post, we looked at how alert noise is rarely accidental. It’s usually the result of sensible decisions layered over time, until responsibility becomes diffuse and response slows. One of the most persistent assumptions behind this pattern is simple. If enough people are notified, someone will take responsibility. After more than fourteen years

Alert Noise Isn’t an Accident — It’s a Design Decision

3 min read In a previous post, The Incident Checklist: Reducing Cognitive Load When It Matters Most, we explored how incidents stop being purely technical problems and become human ones. These are moments where decision-making under pressure and cognitive load matter more than perfect root cause analysis. When systems don’t support people clearly in those moments, teams compensate.

The Incident Checklist: Reducing Cognitive Load When It Matters Most

4 min read In the previous post, we looked at what happens after detection; when incidents stop being purely technical problems and become human ones, with cognitive load as the real constraint. This post assumes that context. The question here is simpler and more practical. What actually helps teams think clearly and act well once things are already

When Things Go Wrong, Systems Should Help Humans — Not Fight Them

3 min read In the previous post, we explored how AI accelerates delivery and compresses the time between change and user impact. As velocity increases, knowing that something has gone wrong before users do becomes a critical capability. But detection is only the beginning. Once alerts fire and dashboards light up, humans still have to interpret what’s happening,

When AI Speeds Up Change, Knowing First Becomes the Constraint

5 min read In a recent post, I argued that AI doesn’t fix weak engineering processes; rather it amplifies them. Strong review practices, clear ownership, and solid fundamentals still matter just as much when code is AI-assisted as when it’s not. That post sparked a follow-up question in the comments that’s worth sitting with: With AI speeding things

Make Your Engineering Processes Resilient. Not Your Opinions About AI

4 min read Why strong reviews, accountability, and monitoring matter more in an AI-assisted world Artificial intelligence has become the latest fault line in software development.  For some teams, it’s an obvious productivity multiplier.  For others, it’s viewed with suspicion.  A source of low-quality code, unreviewable pull requests, and latent production risk. One concern we hear frequently goes

Want to know how much website downtime costs, and the impact it can have on your business?

Find out everything you need to know in our new uptime monitoring whitepaper 2021

*By providing your email address, you agree to our privacy policy and to receive marketing communications from StatusCake.