Welcome!

Machine Learning Authors: Pat Romanski, Yeshim Deniz, Liz McMillan, Elizabeth White, Zakia Bouachraoui

Related Topics: @CloudExpo, Microservices Expo, Containers Expo Blog, Cognitive Computing , Machine Learning , Cloud Security

@CloudExpo: Blog Post

Preparing for Cloud Outages

Your Business Needs a Backup Plan

When businesses think of cloud computing they think of the efficiency, stability and accessibility that it represents. Often the trust in the cloud is so great that it's easy to forget that there should be a backup plan in case an outage occurs. Although cloud providers offer service level agreements, this is not enough to guarantee continuity of service, and businesses still need to prepare for a situation where they experience a cloud outage.

Know the Impacts
Depending on what you've moved to the cloud, the impact of a cloud outage can vary. If your email services or office productivity software is cloud based, the impact can be minimal and simply be some delayed emails or late reports.

On the other hand, if you are completely cloud based, and have your entire IT infrastructure offsite, the impact of a cloud outage would be more severe and could impact your accounting system or sales order processing system, which ultimately could lead to a loss of business revenue.

Know the Escalation Path
Working with your cloud provider and knowing who you can contact for support is important. Make sure your IT staff know whether a ticket needs to be opened for a slowdown in service, or if a call to the network operations center is more appropriate.

Knowing the signs of an outage is critical too. Make sure you have staff trained to recognize the symptoms of a cloud outage and that they know the procedures for a backup plan.

Have a Backup Plan
In the event of a minor cloud outage, knowing how to reroute email, or having alternate productivity software in place can minimize the impact. If you have more critical business services in the cloud, it is a good idea to have a full redundancy plan in place to ensure that your bottom line is not impacted.

At the end of the day, the importance of a disaster recovery plan cannot be understated. No matter how much you trust the cloud, when you take the system out of your hands there will always be an element of risk. Know your vendors service levels, their disaster recovery plans, and your key escalation contacts, but always have a backup solution in place to prevent loss of time, data, and money.

Research for this post was provided by Rand Group - a provider of financial management software for the construction, manufacturing, distribution and oil & gas sectors.

[Image credit: Getbutterfly via Flickr]

More Stories By Tom Caper

Tom Caper is a Sales Manager at Rand Group based out of Houston and has 15+ years experiencing working in the enterprise solutions area (CRM, ERP, etc.). He has a particular interest in software-related technology and B2B sales, social CRM, ERP and cloud computing.

CloudEXPO Stories
With more than 30 Kubernetes solutions in the marketplace, it's tempting to think Kubernetes and the vendor ecosystem has solved the problem of operationalizing containers at scale or of automatically managing the elasticity of the underlying infrastructure that these solutions need to be truly scalable. Far from it. There are at least six major pain points that companies experience when they try to deploy and run Kubernetes in their complex environments. In this presentation, the speaker will detail these pain points and explain how cloud can address them.
The deluge of IoT sensor data collected from connected devices and the powerful AI required to make that data actionable are giving rise to a hybrid ecosystem in which cloud, on-prem and edge processes become interweaved. Attendees will learn how emerging composable infrastructure solutions deliver the adaptive architecture needed to manage this new data reality. Machine learning algorithms can better anticipate data storms and automate resources to support surges, including fully scalable GPU-centric compute for the most data-intensive applications. Hyperconverged systems already in place can be revitalized with vendor-agnostic, PCIe-deployed, disaggregated approach to composable, maximizing the value of previous investments.
When building large, cloud-based applications that operate at a high scale, it's important to maintain a high availability and resilience to failures. In order to do that, you must be tolerant of failures, even in light of failures in other areas of your application. "Fly two mistakes high" is an old adage in the radio control airplane hobby. It means, fly high enough so that if you make a mistake, you can continue flying with room to still make mistakes. In his session at 18th Cloud Expo, Lee Atchison, Principal Cloud Architect and Advocate at New Relic, discussed how this same philosophy can be applied to highly scaled applications, and can dramatically increase your resilience to failure.
Machine learning has taken residence at our cities' cores and now we can finally have "smart cities." Cities are a collection of buildings made to provide the structure and safety necessary for people to function, create and survive. Buildings are a pool of ever-changing performance data from large automated systems such as heating and cooling to the people that live and work within them. Through machine learning, buildings can optimize performance, reduce costs, and improve occupant comfort by sharing information within the building and with outside city infrastructure via real time shared cloud capabilities.
As Cybric's Chief Technology Officer, Mike D. Kail is responsible for the strategic vision and technical direction of the platform. Prior to founding Cybric, Mike was Yahoo's CIO and SVP of Infrastructure, where he led the IT and Data Center functions for the company. He has more than 24 years of IT Operations experience with a focus on highly-scalable architectures.