How to Put Public Sector Data Migration Hassles on the Road to Extinction

A group of people standing in front of a large screen.

2018 AT&T Business Summit: Security “in” and “of” the Cloud

By G C Network | November 10, 2018

While public cloud is undoubtedly an outsized piece of the conversation, news headlines of the latest data breach can make this move a very frightening proposition. The question of how…

A group of people standing in front of a display.

My Brush with Royalty: Queen Latifah

By G C Network | November 9, 2018

Queen Latifah! Hip Hop Icon. Movie Star. Television Star. Fashion Model. Songwriter. Producer. Entrepreneurial Genius!? YES! Dana Elaine Owens, her given name, is co-owner of Flavor Unit Entertainment, a firm that includes…

What’s New in Puppet 5?

By G C Network | November 9, 2018

Puppet 5 is released and comes with several exciting enhancements and features that promise to make configuration management much more streamlined. This article will take a comprehensive look at these…

5 Reasons Why Ansible is the Best CM Tool Out There?

By G C Network | November 9, 2018

Amidst volatile markets, dynamic technology shifts, and ever-increasing customer demands, it is imperative for IT organizations to develop flexible, scalable and high-quality applications that exceed expectations and enhance productivity. A…

Machine learning APIs for Google Cloud Platform

By G C Network | November 9, 2018

Google Cloud Platform (GCP) is considered to be one of the Big 3 cloud platforms among Microsoft Azure and AWS. GCP is widely used cloud solutions supporting AI capabilities to design and…

What Is The Most Important Part of Architecture?

By G C Network | November 9, 2018

I always find it interesting to hear what people view architecture as. A lot of people think it’s just about the design aspect, where you get to put pen to…

Cloud migration best practice Part 4: Executing the migration

By G C Network | September 16, 2018

This series has stepped through cloud migration best practices. After providing an overview, we discussed: Classifying business-critical data. Updating organizational IT governance policies. Application screening and cloud workload selection. How to…

Cloud Migration Best Practice Part 3: Application Portfolio Analysis

By G C Network | September 16, 2018

In part three of this series on cloud migration best practice, I will focus on migrating the application itself. If you haven’t had the opportunity to read our recommendations from part…

Why Use Immutable Storage?

By G C Network | September 11, 2018

Data has become a global currency, and its value has nowhere to go but up.  According to The Economist online, the world’s most valuable resource is no longer oil, but…

CLOUD ACT: What Does That Mean for Your Cloud Storage

By G C Network | August 31, 2018

When Congress names a law after you, it’s getting serious.  That is where we are now with cloud computing.  The Clarifying Lawful Overseas Use of Data Act or CLOUD Act (H.R. 4943) is…


With careful planning and the right technology, Federal, State and Local Government IT Leaders can overcome fears of data migrations, breaking free from archaic procedures to lead the pack 
By David Wegman, Vision Solutions, Senior Vice President, Integrated Accounts


Jurassic World, the latest installment in the Jurassic Park film series, opened this week – and there’s a lot of hype surrounding the premiere as fans immerse themselves in a world of Mesozoic Era-inspired fantasy. While the creatures that make the theme park their home are strikingly realistic, their real-life counterparts became extinct millennia ago.  Many believe that the once mighty dinosaur population fell in large part because it failed to evolve with the changing world around it. Public sector institutions face a similar plight today, especially as technology advancements demand they constantly evolve in order to keep up.
Much like the dinosaurs fought for survival, governmental organizations must fight for resources. They must embrace change in order to thrive, and part of that involves modernizing systems, streamlining processes and migrating vast amounts of data. However, many organizations postpone such work due to uncertainties about the impact and technology risks associated with these procedures, including the inherent downtime associated with most migration methodologies.
Many public sector CIOs and IT leaders are concerned about the fallout from failed migrations, which are a painful waste of time and resources. And their concerns are not unfounded: In its 2015 State of Resilience report, Vision Solutionsrevealed that 36 percent of respondents had experienced a migration failure. While failures are a relatively common occurrence, they are not inevitable. A thorough planning process and the right resources go a long way to improve the chances of success.
Regardless of the reason for a migration, significant complexity and potential pitfalls litter the path from point A to point B. In addition to understanding the migration process and identifying who is going to do the work, users must assess downtime’s potentially negative impacts.
The fact is, migrations are complex; even those that sound simple are inherently complex. Because most servers and databases are not single instances within a data center but interconnected to other systems and databases, including mid-range and mainframe systems, there is immense variety that complicates migrations today. These various platforms need to be coordinated in migration waves to mitigate their impacts.  This all presents complexity – and complexity presents risks.
The first task toward successful migrations is to map out a thorough migration plan beforehand: IT needs to determine what they are going to migrate, what it is all connected to, who is going to do the work and when they can get a “migration window” from their unit, as well as how they will navigate around the many real-time issues that may arise along the way. Planning ahead for potential issues gives IT clarity on factors that may affect the migration process, allowing the migration team to address problems in advance and in real-time.
All migrations need testing before deployment; and testing further contributes to the time and resource-intensive nature of migrations. Traditional pre-migration testing can take anywhere from a couple of weeks to a couple of months, depending on how complex the applications, databases and server inter-connections are in the data center. The process typically involves halting production periodically to take snapshots of data and testing those snapshots.  Each time the process is completed, IT must restore the database and start over. This typically involves multiple test runs and multiple cycles within the organization. The entire process can take anywhere from hours to entire days depending largely on the factors outlined above.
IT often faces an uphill battle in convincing leadership to agree to a migration project due to downtime risks and impacts. This can cause substantial delays, compounding the migration’s complexity. Elected and appointed officials may not always have exact numbers on hand, but they do realize that downtime is costly. Nineteen percent of respondents in the Vision Solutions 2015 State of Resilience report indicated that the cost of downtime ranged from $10,000 to $50,000 per hour. Fearing the steep costs of downtime and associated risks, leadership may hesitate to green light migration projects, preventing successful execution.
How can public sector IT leaders address this problem? Government IT leaders need more than ever to seek out trusted partners with a track record of helping organizations like them accomplish important migrations and system upgrades. Rather than reacting to perceived risks, limited expertise or cost of downtime, they should be confident and proactive, following best practices to set them up for success. Organizations that embrace proven technology and methodologies will be well-positioned to realize the full benefits of smooth migrations.
Uncertainty, risk and extended downtime don’t need to be migration realities. By working with a trusted partner and utilizing modern technology and methodology, public sector IT leaders can achieve near-zero downtime during migrations, minimizing impact on the organization and users. But to do so, they should consider the following when selecting a migration solution:
1. Real-time replication is paramount: Organizations should look for solutions that offer the most flexibility and currency of data possible while minimizing impact to users during testing and migration. This typically requires a software-based solution that replicates any activity taking place on the production server to the target server in real-time, allowing IT to keep the production server up and running rather than freezing it or periodically pausing it for snapshots. The production server remains fully functional, data is as current as the last transaction and users continue working. IT can test applications on the new server, and prove the migration methodology and plan, without impacting the production environment. Ultimately, this makes IT more productive on other tasks with improved uptime – all while migration is taking place.
A second consideration is how to take the distance from production server to target server out of the equation. Because real-time replication sends the changes as they occur, it minimizes the amount of communication line and distance becomes less of an issue. When coupled with compression and throttling in a product, this creates a high degree of efficiency.
Finally, because databases and servers are maintained in sync at all times, IT does not need to freeze production and wait for final validation of the testing server to finally perform the migration. Weekend migrations are no longer the norm as the switch to new environments can occur at any time the organization is ready and take place in as little as 20 minutes– a notable improvement over switch times in traditional migrations.
2. Unified consoles simplify the process: Another feature government IT leaders should demand in their migration solution is a unified console that allows IT to work on all types of migrations with a common workflow across operating systems and platforms. This provides a major advantage as it mitigates the need for different skillsets typically required for different types of migrations by platform or workload.
While IT staff certainly needs to understand the underlying architectures and databases, by using a uniform console and workflow, it reduces training time and maximizes the existing team’s skillset. A single operator can perform parallel migrations across multiple platforms after product training sessions, minimizing the drain on resources.
3. Consolidating migration streams delivers faster execution: Simultaneous executions also ease the impacts to the organization. A solution that allows users to run parallel streams of migrations saves organizations significantly more time than traditional methods. This method facilitates near-zero downtime, shortens time to completion, mitigates costs and frees up IT resources to focus on other projects.
4. Automation minimizes risk: Traditional methods typically require a fair degree of manual work, which equals a higher degree of risk.  While no migration can happen without people, solutions that reduce the required amount of human interaction from migrations via automation diminish risk. This is very important for organizations to keep in mind, as failing to choose a solution that provides APIs the ability to automate as much of the work as possible introduces additional human interaction and therefore risk. This lack of automation often results in failed migrations, migrations that run over budget or last longer than expected.
5. Hardware- and software-independent solutions enable flexibility: Every server is different and topologies change rapidly. Migrating across server types, chipsets, storage devices, databases, versions and the like all need to be addressed in a migration plan. A hardware- and software-independent solution reduces the risk potential in these areas. This model allows users to migrate data seamlessly from any one type of environment to another. The options are virtually endless – from physical to virtual to cloud across any operating system, chipset or storage device.
Using platform-independent technology makes many scenarios possible including migrating between storage from different vendors, migrating to a server located anywhere in the world, consolidating servers with many-to-one migration and moving operations to a new data center across extended distances with very little downtime.
Evolve Continuously to Achieve Migration Success
While data migrations will always entail a certain amount of risk and downtime, modern solutions have greatly improved chances for a positive outcome. IT leaders at Federal, State and Local governments and those at their agencies that act confidently, instead of out of fear, to take on important migrations for their organization will come out on the top of the food chain, evolving to thrive to the benefit of both their organization and the taxpayer.  

 

( This content is being syndicated through multiple channels. The opinions expressed are solely those of the author and do not represent the views of GovCloud Network, GovCloud Network Partners or any other corporation or organization.)

Cloud Musings

( Thank you. If you enjoyed this article, get free updates by email or RSS – © Copyright Kevin L. Jackson 2015)

Follow me at https://Twitter.com/Kevin_Jackson
Posted in

G C Network