Skip to content Skip to footer

What to consider before buying cloud downtime insurance

Florian Malecki

Florian Malecki

EVP of Marketing at Arcserve
Image credit: JLStock / Shutterstock.com

As cloud computing becomes ubiquitous, more companies are exposed to incidents that cause downtime, which can be disastrous.

According to Gartner, the average cost of IT downtime is a staggering $5,600 per minute. And then there are those additional costs that don’t necessarily show up as monetary losses, such as the cost of an interruption that pulls IT people away from their regular work to get your company back up and running.

It is one reason why cloud downtime insurance has taken off in recent years. Downtime insurance providers cover clients for short-term cloud outages, network crashes, and platform failures that last up to 24 hours.

They happen often. Cloud insurance provider Parametrix says that, on average, one of the three major public cloud providers – Microsoft Azure, AWS, and Google Cloud – has an outage lasting at least 30 minutes every three weeks.

Cloud downtime insurance can be a helpful safety net for businesses, but it is not a complete solution. It’s important to remember that this kind of insurance can’t guarantee that your business remains in operation during a period of downtime.

Yes, the insurance will cover you for any short-term losses you incur. But it will not cover the loss of goodwill, damage to your brand image, and loss of customer loyalty when your business can’t deliver.

Instead of relying 100% on cloud downtime insurance, organisations should pursue these three strategies to weather cloud downtime and other unexpected events.

1: Have a sound recovery plan

Think your data is safe and secure when you move it to a cloud provider? Think again. Last year, a fire at the data centre of French web hosting service OVHcloud (Europe’s largest cloud provider) caused the loss of massive amounts of customer data. It impacted government agencies, e-commerce companies, and banks, among others.

Backing up your data to the cloud or on-premise is a critical and cost-effective first step in any disaster recovery plan. But it’s only the first step. It would help if you also had a plan to quickly recover your data in an emergency. Think of your business journey as a trip on a cruise ship. Just as a cruise ship regularly tests its lifeboats (weekly, in case you’re wondering), you should test your recovery plan often. You should simulate disruptions and see how well your recovery plan works. You should also regularly test your backup images and fix any problems. Your recovery plan is your lifeboat.

2: Implement your backup and recovery solution

Cloud security is not solely the responsibility of your cloud provider. It’s your responsibility as well. Cloud providers usually promise to secure their infrastructure and services. But securing operating systems, platforms, and data – that’s on you.

Cloud providers will not guarantee the safety of your data. No matter what cloud platform you use, the data is still owned by you, not the provider. Many cloud providers recommend that their customers use third-party software to protect their data.

You can comprehensively secure your data with a reliable cloud backup and recovery solution. You can also get the control you need. You should implement a cloud backup and recovery solution that protects your data by automatically backing up your information every 15 minutes and gives you multiple points of recovery. This guarantees that your data is continuously protected while providing quick access and visibility to it 24/7.

3: Be proactive: be data resilient

A lot of companies don’t test their data recovery plans. Many don’t even have a recovery plan. Don’t be like them. Have a recovery plan and test it often. Be proactive, not reactive. Be data resilient.

A data resilience strategy ensures business continuity in the event of a disruption. It is built on recovery point objectives (RPOs) and recovery time objectives (RTOs), and you should regularly test to guarantee that the RPOs and RTOs can be achieved.

Your RPO determines your backup frequency. In essence, it’s your tolerance for data loss. Some organisations can tolerate a data loss of 24 hours, so they back up their data every 24 hours. Their RPO is 24. Other organisations, such as those in finance and healthcare, absolutely cannot tolerate a data loss of 24 hours. Their RPOs are set to milliseconds.

Your RTO measures the downtime you can accept between a data loss and recovery. It’s how long you can be down before your business incurs severe damages. Your RTO determines your disaster recovery plan investment. If your RTO is one hour, you need to invest in solutions that get you back up and running within an hour.

Establishing your RPO and RTO and then implementing the solutions you need to achieve them are the keys to data resilience.

Final takeaway

We live in a world of growing cybersecurity threats, more frequent natural disasters, and black swan events arriving in flocks. Every day, organisations are brought to their knees out of the blue. That’s why more of them are purchasing cloud downtime insurance. But it is critical to realise that this type of insurance alone does not constitute a data protection plan. It is best viewed as a complement to your backup and recovery efforts. Never consider it a replacement.

Show CommentsClose Comments

Leave a comment