Story image

How to avoid that next storage hardware refresh

11 Apr 17

Customers tell me repeatedly that they dread the phone call from their hardware storage vendor. Every three or four years, the call is predictable:

“Great news John, due to economies of scale the cost per GB of storage has dropped and performance has increased – we can also provide you with some flash to help improve performance. Plus, we have extra options like deduplication, encryption and replication – we will need to charge you for these.”

Wait, there is more: “Yes, all you need to do is replace all that perfectly good working storage with our new array, and due to your explosive growth it will cost you slightly more than your last refresh. Our professional services team can help you with a migration project to move data from our old frame to our new frame.”

Sound familiar? How would you like to change the conversation?

“That is OK Bill, you know how we virtualised our servers all those years ago, and it changed how we purchased servers?"

“Well we are doing the same to our storage, and now we no longer need that hardware upgrade. In fact we are adding 1TB of RAM for about A$12,000, 5TB of high performance flash for about $A10,000.

For extra capacity, we are using multiple 10TB drives – they are costing us only about $A800 each and we can interconnect all our existing disparate SANs and NAS’s into a single storage pool by using Objective-Defined Storage.

Objective-Defined Storage manages in real-time data tiering between hot, warm and cold data to free up RAM and SSD when an application no longer needs those expensive resources.

All resources can be shared across all applications – even applications on different hosts. It can even create up to eight resilient live instances of critical data using our existing storage including these inexpensive high capacity drives – pretty cool. Oh, and thank you for your suggestion, but we will give it a miss.”

Software is king! And your data is your kingdom.

Why are you replacing your perfectly good storage?  Reliability? Performance? Scalability?

Using Objective-Defined Storage, which is a software-only solution that allows users to extend the life of existing storage assets, exponentially increase performance and deliver totally elastic scale-up, scale-out, scale anyway you require, including the cloud capacity.

The technology extends the life of existing storage for a couple of years, allowing users to avoid those expensive fork-lift upgrades.

Using Objective-Defined Storage, typically allows users to free up between 50 to 80 per cent of capacity of a SAN or NAS by moving cold data to high capacity low cost devices, or the cloud; deliver exponential performance improvement using commodity SSDs; and improve data availability.

Two of the major outcomes are finally eliminating cost and complexity from a storage ecosystem by deploying the storage solution for the new enterprise.

Just like virtualisation changed the way we use and deploy servers, storage virtualisation de-couples data from the underlying storage hardware and allows data to move freely and ubiquitously across all storage, regardless of vendor, make or model.

This delivers data freedom to a business and puts an end to vendor lock-in, storage silos and forklift upgrades once and for all.

Article by Greg Wyman, Vice President Asia Pacific, ioFabric

Is Supermicro innocent? 3rd party test finds no malicious hardware
One of the larger scandals within IT circles took place this year with Bloomberg firing shots at Supermicro - now Supermicro is firing back.
Record revenues from servers selling like hot cakes
The relentless demand for data has resulted in another robust quarter for the global server market with impressive growth.
Opinion: Critical data centre operations is just like F1
Schneider's David Gentry believes critical data centre operations share many parallels to a formula 1 race car team.
MulteFire announces industrial IoT network specification
The specification aims to deliver robust wireless network capabilities for Industrial IoT and enterprises.
Google Cloud, Palo Alto Networks extend partnership
Google Cloud and Palo Alto Networks have extended their partnership to include more security features and customer support for all major public clouds.
DigiCert conquers Google's distrust of Symantec certs
“This could have been an extremely disruptive event to online commerce," comments DigiCert CEO John Merrill. 
Schneider Electric's bets for the 2019 data centre industry
From IT and telco merging to the renaissance of liquid cooling, here are the company's top predictions for the year ahead.
China to usurp Europe in becoming AI research world leader
A new study has found China is outpacing Europe and the US in terms of AI research output and growth.