|By Jamal Mazhar||
|November 30, 2012 03:00 PM EST||
Use of cloud for DR solutions is becoming more common, even the organizations which are not using cloud for mission critical production applications are moving towards using cloud for application DR.
Business Case for Using Cloud for the DR
Faster Recovery Time Objective (RTO): Typically DR requires lengthy manual processes to fully restore the business applications at the DR site. Having backup data and servers at the DR site is easy, however, restoring the entire application or service takes time. E.g. full application restoration requires starting services in specified order, performing dns and other configuration updates etc. In Cloud, the IaaS APIs provide ability to use automation solutions like Kaavo IMOD to fully restore the business applications automatically without manual intervention. As a result organizations get predictable recovery and reduced RTO. Automating the service or application recovery can reduce RTO to minutes from hours or days.
Shorter Recovery Point Objective (RPO): Instead of relying on offsite tape backups, organizations can reduce their RPO to minutes by maintaining near real-time data backups in the Cloud. For faster transfer of large data dedicated lines can be established between the customer datacenters and the cloud. The cost of the dedicated line depends on the distance of the customer datacenter from the cloud providers' peering point. For most use cases VPN lines over internet are sufficient for transferring data between customer datacenter and the cloud.
Lower Costs: Typically organizations pay high price for standby infrastructure, especially servers at the DR site. Using cloud there is no need to pay for the servers when they are not in use at the DR site. Pay as you use infrastructure model significantly reduces DR costs without compromising the service levels.
Following are some of the best practices and lessons learned from the Cloud DR solutions we have implemented so far:
Cloud DR is Different than Traditional DR
Unlike traditional DR solutions which relies on having a backup infrastructure for the entire datacenter requiring large and costly implementation, Cloud DR can be implemented incrementally application by application. For example it is common for organizations to have a large shared database with multiple schemas supporting various applications. In majority of cases this sharing is driven by server consolidation to increase the utilization of internal infrastructure. Not all applications using a shared database have same service level requirements. Some applications are more critical than others, so as long as schemas and application data is different, it is better to remove the dependency on shared database by having the right size database for each application in the cloud. This allows optimal prioritization and incremental delivery of the DR project based on the service levels of the individual applications.
Migration of Applications Using Single Sign-on with LDAP
When planning DR for individual applications it is important to identify the dependent services and making sure that the dependent services would be available as a part of the DR solution. Enterprise customers typically use Single Sign-on with LDAP for managing authentication. So best practice is to treat the Single Sign-on Service as the critical application and implement the DR solution for bringing up the Single Sign-on Service first during the DR process. An automation solution like Kaavo IMOD enables customers to restore applications and services in the specified order automatically during DR without any manual intervention. During a real DR scenario there are many things going and it is easy to make mistakes under pressure if the application restoration process is not fully automated. To prevent surprises during actual DR, it is important to have a fully automated solution for restoring applications and services.
Restoring Back to Normal Operations after DR
This is one area which is often overlooked or under planned in DR projects. For companies using their own datacenters for production applications and using cloud for DR, processes and automation must be implemented to fully restore the applications in the customer production datacenter using the latest data from the cloud DR once the primary datacenter is back online. This step is not required for applications which are using cloud as their primary site. E.g. if an application is running in one cloud zone and after DR it is running in a different cloud zone there is no need to restore it back to the first cloud zone as long as service levels for both cloud zones are same. If you are deploying new applications it best to design for failure. E.g. a distributed application running across various regions and cloud providers eliminate the need for traditional DR planning for the application as handling of failure of individual components is built in the design and deployment model of the application.
Handling Compliance in Cloud, e.g., HIPAA, PCI, SOX, SAS-70 etc.
Using available security technologies and processes several companies have implemented applications in the cloud compliant to various compliance standards, e.g. HIPAA, PCI, SOX, SAS-70 etc. Each compliance standard has its own nuances; basically with proper planning you can address all compliance related issues. This is a big topic on its own so please contact us if you have specific questions about this. Cloud providers have published various case studies and best practices, e.g. white paper by Amazon on HIPAA compliance.
Handling Public and Private DNS
A common use case for enterprise applications is to have a public DNS for public access and a private DNS over internal network for accessing the backend services and databases etc. In these situations it is best to use virtual private cloud like AWS VPC or to overlay a private network with the same IP address range as internal datacenter on any public cloud using Open Source solutions (refer to this blog - Building a Private Cloud within a Public Cloud for details on how to implement a secure private network on any public cloud). For updating the public DNS entries for the restored application in the cloud we use DNS automation services like AWS Route 53 or EasyDNS. Leveraging these services, Kaavo IMOD automatically updates the Public DNS for the applications as a part of the restoration during DR.
Keeping Application Database Up-To-Date
It is common for applications to have large databases. Moving the data to the cloud and keeping it current requires first loading the entire database in cloud and then sending and merging incremental data to the database in the cloud. To address this use case instead of maintaining a hot backup we use Kaavo IMOD to automatically bring up the database servers in cloud whenever the new incremental backup is available and merge the incremental backup then save the merged database and shutdown the servers in the cloud. This way in case of DR we always have the latest merged database available for restoring the application. This approach provides reasonable RTO without incurring the additional costs of maintaining a hot database backup.
Applying and Maintaining Patches
A typical application requires following two types of updates during its lifecycle:
Updating Application Code: This is quite easy as using Kaavo IMOD we setup automation to pick up the latest code and configuration for the application from the production deployment. This automation ensures that the application code and configuration changes for the new release of the application or service are available in the cloud for the DR.
OS Patches and Third-Party Software Updates: Sometimes custom patches or updates to third party software or OS are required. For these types of changes it is best to include them as a part of change control process requiring sign-off from the team owning the DR process. The DR team can review the change and if required make and test the needed changes to DR automation for the application.
Large scale deployments present unique planning challenges, system commissioning hurdles between IT and OT and demand careful system hand-off orchestration. In his session at @ThingsExpo, Jeff Smith, Senior Director and a founding member of Incenergy, will discuss some of the key tactics to ensure delivery success based on his experience of the last two years deploying Industrial IoT systems across four continents.
Jul. 25, 2016 06:00 AM EDT Reads: 1,454
There will be new vendors providing applications, middleware, and connected devices to support the thriving IoT ecosystem. This essentially means that electronic device manufacturers will also be in the software business. Many will be new to building embedded software or robust software. This creates an increased importance on software quality, particularly within the Industrial Internet of Things where business-critical applications are becoming dependent on products controlled by software. Qua...
Jul. 25, 2016 05:45 AM EDT Reads: 1,351
SYS-CON Events has announced today that Roger Strukhoff has been named conference chair of Cloud Expo and @ThingsExpo 2016 Silicon Valley. The 19th Cloud Expo and 6th @ThingsExpo will take place on November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. "The Internet of Things brings trillions of dollars of opportunity to developers and enterprise IT, no matter how you measure it," stated Roger Strukhoff. "More importantly, it leverages the power of devices and the Interne...
Jul. 25, 2016 05:00 AM EDT Reads: 2,007
Machine Learning helps make complex systems more efficient. By applying advanced Machine Learning techniques such as Cognitive Fingerprinting, wind project operators can utilize these tools to learn from collected data, detect regular patterns, and optimize their own operations. In his session at 18th Cloud Expo, Stuart Gillen, Director of Business Development at SparkCognition, discussed how research has demonstrated the value of Machine Learning in delivering next generation analytics to imp...
Jul. 25, 2016 04:15 AM EDT Reads: 2,395
In addition to all the benefits, IoT is also bringing new kind of customer experience challenges - cars that unlock themselves, thermostats turning houses into saunas and baby video monitors broadcasting over the internet. This list can only increase because while IoT services should be intuitive and simple to use, the delivery ecosystem is a myriad of potential problems as IoT explodes complexity. So finding a performance issue is like finding the proverbial needle in the haystack.
Jul. 25, 2016 03:45 AM EDT Reads: 2,166
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform. In his session at @ThingsExpo, Craig Sproule, CEO of Metavine, demonstrated how to move beyond today's coding paradigm and shared the must-have mindsets for removing complexity from the develo...
Jul. 25, 2016 01:45 AM EDT Reads: 1,221
SYS-CON Events announced today that MangoApps will exhibit at the 19th International Cloud Expo, which will take place on November 1–3, 2016, at the Santa Clara Convention Center in Santa Clara, CA. MangoApps provides modern company intranets and team collaboration software, allowing workers to stay connected and productive from anywhere in the world and from any device.
Jul. 25, 2016 01:30 AM EDT Reads: 1,254
Basho Technologies has announced the latest release of Basho Riak TS, version 1.3. Riak TS is an enterprise-grade NoSQL database optimized for Internet of Things (IoT). The open source version enables developers to download the software for free and use it in production as well as make contributions to the code and develop applications around Riak TS. Enhancements to Riak TS make it quick, easy and cost-effective to spin up an instance to test new ideas and build IoT applications. In addition to...
Jul. 25, 2016 12:30 AM EDT Reads: 1,878
The 19th International Cloud Expo has announced that its Call for Papers is open. Cloud Expo, to be held November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, brings together Cloud Computing, Big Data, Internet of Things, DevOps, Digital Transformation, Microservices and WebRTC to one location. With cloud computing driving a higher percentage of enterprise IT budgets every year, it becomes increasingly important to plant your flag in this fast-expanding business opportuni...
Jul. 25, 2016 12:15 AM EDT Reads: 2,493
"We've discovered that after shows 80% if leads that people get, 80% of the conversations end up on the show floor, meaning people forget about it, people forget who they talk to, people forget that there are actual business opportunities to be had here so we try to help out and keep the conversations going," explained Jeff Mesnik, Founder and President of ContentMX, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 24, 2016 11:30 PM EDT Reads: 1,276
“delaPlex Software provides software outsourcing services. We have a hybrid model where we have onshore developers and project managers that we can place anywhere in the U.S. or in Europe,” explained Manish Sachdeva, CEO at delaPlex Software, in this SYS-CON.tv interview at @ThingsExpo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 24, 2016 11:00 PM EDT Reads: 1,511
From wearable activity trackers to fantasy e-sports, data and technology are transforming the way athletes train for the game and fans engage with their teams. In his session at @ThingsExpo, will present key data findings from leading sports organizations San Francisco 49ers, Orlando Magic NBA team. By utilizing data analytics these sports orgs have recognized new revenue streams, doubled its fan base and streamlined costs at its stadiums. John Paul is the CEO and Founder of VenueNext. Prior ...
Jul. 24, 2016 10:45 PM EDT Reads: 1,990
IoT is rapidly changing the way enterprises are using data to improve business decision-making. In order to derive business value, organizations must unlock insights from the data gathered and then act on these. In their session at @ThingsExpo, Eric Hoffman, Vice President at EastBanc Technologies, and Peter Shashkin, Head of Development Department at EastBanc Technologies, discussed how one organization leveraged IoT, cloud technology and data analysis to improve customer experiences and effi...
Jul. 24, 2016 10:00 PM EDT Reads: 1,966
The Internet of Things will challenge the status quo of how IT and development organizations operate. Or will it? Certainly the fog layer of IoT requires special insights about data ontology, security and transactional integrity. But the developmental challenges are the same: People, Process and Platform and how we integrate our thinking to solve complicated problems. In his session at 19th Cloud Expo, Craig Sproule, CEO of Metavine, will demonstrate how to move beyond today's coding paradigm ...
Jul. 24, 2016 09:45 PM EDT Reads: 2,128
Internet of @ThingsExpo, taking place November 1-3, 2016, at the Santa Clara Convention Center in Santa Clara, CA, is co-located with the 19th International Cloud Expo and will feature technical sessions from a rock star conference faculty and the leading industry players in the world and ThingsExpo Silicon Valley Call for Papers is now open.
Jul. 24, 2016 09:00 PM EDT Reads: 2,467
Big Data engines are powering a lot of service businesses right now. Data is collected from users from wearable technologies, web behaviors, purchase behavior as well as several arbitrary data points we’d never think of. The demand for faster and bigger engines to crunch and serve up the data to services is growing exponentially. You see a LOT of correlation between “Cloud” and “Big Data” but on Big Data and “Hybrid,” where hybrid hosting is the sanest approach to the Big Data Infrastructure pro...
Jul. 24, 2016 07:45 PM EDT Reads: 1,860
A critical component of any IoT project is what to do with all the data being generated. This data needs to be captured, processed, structured, and stored in a way to facilitate different kinds of queries. Traditional data warehouse and analytical systems are mature technologies that can be used to handle certain kinds of queries, but they are not always well suited to many problems, particularly when there is a need for real-time insights.
Jul. 24, 2016 07:30 PM EDT Reads: 1,701
"My role is working with customers, helping them go through this digital transformation. I spend a lot of time talking to banks, big industries, manufacturers working through how they are integrating and transforming their IT platforms and moving them forward," explained William Morrish, General Manager Product Sales at Interoute, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 24, 2016 07:30 PM EDT Reads: 2,064
With 15% of enterprises adopting a hybrid IT strategy, you need to set a plan to integrate hybrid cloud throughout your infrastructure. In his session at 18th Cloud Expo, Steven Dreher, Director of Solutions Architecture at Green House Data, discussed how to plan for shifting resource requirements, overcome challenges, and implement hybrid IT alongside your existing data center assets. Highlights included anticipating workload, cost and resource calculations, integrating services on both sides...
Jul. 24, 2016 07:00 PM EDT Reads: 1,933
"We are a well-established player in the application life cycle management market and we also have a very strong version control product," stated Flint Brenton, CEO of CollabNet,, in this SYS-CON.tv interview at 18th Cloud Expo, held June 7-9, 2016, at the Javits Center in New York City, NY.
Jul. 24, 2016 06:45 PM EDT Reads: 1,767