UPDATED 12:49 EDT / JANUARY 09 2013

Storage Migration Costs Can Exceed Array Prices

In the latest in his series of Wikibon Professional Alerts on Software-led Infrastructure, “Moving to Virtualized Clustered Storage”, Wikibon CTO David Floyer examines the full cost of storage array replacement and discovers that the costs involved in data migration can exceed the price of the original array itself. This expense is often regarded as a depreciated staffing or sunk cost and is therefore not managed as part of the cost of replacing an old array. But, he says, the cost and complexity of moving data results in leaving data on suboptimal arrays an can delay the move to software-led storage.

Migration costs include:

  • The overlapping cost of the old array (five months of a three-year lease),
  • The operational cost of planning and executing the migration of volumes from the old array to the new one, and,
  • The cost of buying the new array five months earlier than required, based on an annual decline in storage costs of about 35% per year.

These costs, Floyer’s analysis shows, typically amount to 45% of the total cost of ownership of the array, presuming that the migration can be completed in five months. If, as sometimes happens, the migration takes a full year, the cost can exceed the cost of the original array.

The complexity comes from moving the live data without shutting down the associated application. These problems are a major inhibitor to providing a scale-out, software-led infrastructure.

Vendors have not drawn attention to these costs and issues because they will cause a negative marketing perception. However, that does not mean they are not aware that this is a problem that needs fixing. Various major vendors, including IBM (SAN Volume Controller), Hitachi (High Availability Manager), EMC (VPLEX), HP (3PAR), and NetApp (ONTAP 8.1), have fielded federated storage systems that potentially can cut the cost and decrease the complexity, providing a technical foundation for moving to software-led storage.

Of these, Floyer says NetApp ONTAP 8.1 is the strongest technology. The present version only works inside a single data center, but he predicts that it will be extended to work across Metro distances and to include data storage on solid state devices in servers in future releases.

His recommendations are for CIOs, CTOs, and senior storage specialists to push their storage vendors to lay out a clear pathway to scale-out architectures, equipment, and software that will enable data-as-a-service.

As with all Wikibon research, this Alert is available in its entirety on the public Wikibon Web site. IT professionals are invited to register for membership in the Wikibon community. This allows them to comment on research and publish their own Professional Alerts, tips, questions, and relevant white papers. It also subscribes them to invitations to the periodic Peer Incite meetings, at which their peers discuss the solutions they have found to real-world problems, and to the Peer Incite Newsletter, in which Wikibon and outside experts analyze aspects of the subjects discussed in these meetings.


A message from John Furrier, co-founder of SiliconANGLE:

Your vote of support is important to us and it helps us keep the content FREE.

One click below supports our mission to provide free, deep, and relevant content.  

Join our community on YouTube

Join the community that includes more than 15,000 #CubeAlumni experts, including Amazon.com CEO Andy Jassy, Dell Technologies founder and CEO Michael Dell, Intel CEO Pat Gelsinger, and many more luminaries and experts.

“TheCUBE is an important partner to the industry. You guys really are a part of our events and we really appreciate you coming and I know people appreciate the content you create as well” – Andy Jassy

THANK YOU