DCIG 2019-20 Enterprise Deduplication Backup Target Appliance Buyer’s Guide Now Available

DCIG is pleased to announce the availability of its 2019-20 Enterprise Deduplication Backup Target Appliance Buyer’s Guide which helps enterprises assess the enter­prise deduplication backup target appliance marketplace and identify which appliance may be the best fit for their environment. This Buyer’s Guide includes data sheets for 19 enterprise deduplication backup target appliances that achieved rankings of Recommended and Excellent. These products are available from five vendors including Cohesity, Dell EMC, ExaGrid, HPE, and NEC.

Enterprises rarely want to talk about the make-up of the infrastructure of their data centers anymore. They prefer to talk about artificial intelligence, cloud adoption, data analytics, machine learning, software-defined data centers, and uninterrupted business operations. As part of those discussions, they want to leverage current technologies to drive new insights into their business and, ultimately, create new opportunities for busi­ness growth or cost savings because their underlying data center technologies work as expected.

The operative phrase here becomes “works as expected”, especially as it relates to Enterprise Deduplication Backup Target Appliances. Expectations as to the exact features that an enterprise deduplication backup target appliance should deliver can vary widely.

If an enterprise only wants an enterprise deduplication backup target appliance that meets traditional data center requirements, every appliance covered in this Buyer’s Guide satisfies those needs. Each one can:

  • Serve as a target for backup software.
  • Analyze and break apart data in backup streams to optimize deduplication ratios.
  • Replicate backup data to other sites
  • Replicate data to the cloud for archive, disaster recovery, and long-term data retention.

While the appliances from each provider uses different techniques to accomplish these objectives and some perform these tasks better than others depending on the use case, each one does deliver on these objectives.

But for enterprises looking for a solution that enables them to meet their broader, more strategic objectives, only a couple of providers covered in this Buyer’s Guide, appear to be taking the appropriate steps to position enterprises for the software-defined hybrid data center of the future. Appliances from these provid­ers better position enterprises to perform next generation data lifecycle management tasks while still providing enterprises with the necessary features to accomplish traditional backup and recovery tasks.

It is in this context that DCIG presents its DCIG 2019-20 Enterprise Deduplication Backup Target Appliance Buyer’s Guide. As in the development of all prior DCIG Buyer’s Guides, DCIG has already done the heavy lifting for enterprise technology buyers by:

  • Identifying a common technology need with competing solutions
  • Scanning the environment to identify available products in the marketplace
  • Gathering normalized data about the features each product supports
  • Providing an objective, third-party evaluation of those features from an end-user perspective
  • Describing key product considerations and important changes in the marketplace
  • Presenting DCIG’s opinions and product feature data in a way that facilitates the rapid comparisons of various products and product features

The products that DCIG ranks as Recommended in this Guide are as follows (in alphabetical order):

Access to this Buyer’s Guide edition is available immediately by following this link to any of the following DCIG partner sites:

TechTrove

HPE




Analytics, Microservices and Scalable Storage Finding Their Way onto Backup Appliances

Companies of all sizes pay more attention to their backup and recovery infrastructure than perhaps ever before. While they still rightfully prioritize their production infrastructure over their backup one, companies seem to recognize and understand that can use backups as more than just insurance policies to recover their production data. This is resulting in cutting edge innovations such as analytics, microservices, and scalable storage finding their way into backup solutions in general and backup appliances specifically.

One of the challenges associated with innovative technologies finding their way into data protection solutions stems from cost. Companies tend to see backup as a cost of doing business and less of an investment in the company’s future. This viewpoint results in them only spending what they must on backup.

Since new technologies tend to be more costly, this tends to inhibit the introduction and adoption of new technologies on the backup appliances. More than one backup provider has told me in briefings with me they would gladly introduce more innovative technologies into their backup solutions. There simply has not been the budget-backed demand for them.

For these reasons, backup may never become a maelstrom of innovation. Nevertheless, changes in how companies want to manage and use their backup infrastructure is driving the introduction of some of the latest and greatest technologies into backup solutions. Here are three that DCIG has observed finding their way onto backup appliances.

  1. Analytics. Backup is ripe for picking when it comes to applying analytics such as artificial intelligence and machine learning technologies to the backup process. Companies keep multiple copies of data in a single data store for multiple years. This would lead one to believe there is more value than just data recovery that companies can glean from their backup data repositories … if they can just figure out what that value prop is.

As it turns out, cybercriminals might be the ones who do the most to help companies derive more value from their backup data stores. As cybercriminals get more sophisticated in their malware attacks, companies are turning to backups to help them detect the presence of malware in their production environment by analyzing backup data.

Asigra incorporates the use of cybersecurity software in its Cloud Backup software. This software can analyze data as it is backed up or recovered for the presence of malware. Unitrends also uses analytics to detect for the presence of ransomware in backups. It compares changes between backups and looks for unusual or unexpected activity in data between backups.

  1. Microservices. Using microservices sto accelerate application development has been one of the hottest topics in technology for the last few years. Here again, microservices have been slow to gain a foothold in data protection though StorageCraft on its OneXafe solution represents one of the first to find a practical application for microservices.

StorageCraft gives companies multiple ways to use their OneXafe appliance. If they use SSDs in its all-flash 5410 model, companies can use it for production file services. But OneXafe was originally designed for use as a secondary storage appliance in the form of a deduplication backup target.

The underlying software platform upon which StorageCraft built OneXafe lends itself well to running data protection microservices in Docker containers. Companies may enable this data protection service on the OneXafe platform at any time. Using OneXafe’s architecture, its ShadowXafe data protection service has access to all the storage across all the nodes in a OneXafe deployment without the design and storage limitations inherent in competing products.

  1. Scalable storage. Scalable storage represents the trend being most rapidly adopted by data protection solutions. Companies struggle to forecast data growth in their production environment. Having to manage the corresponding storage growth in their backup environment only adds to their headaches.

More backup appliances than ever give companies the flexibility they need to easily scale their storage without the corresponding headaches that they have faced in the past. Some examples of this innovation include:

  • Asigra has partnered with iXsystems to deliver the Asigra TrueNAS appliance that scales to over 10PB.
  • StorageCraft merged with ExaBlox and now offers its OneXafe appliance which likewise can scale up to over a petabyte in a single logical configuration.
  • Others like HYCU have partnered with Nutanix to deliver this scalable storage functionality.

Very few IT professionals choose IT out of a desire to merely “feed the beast.” Most want to innovate and adopt new and exciting technologies that create new value for their organizations.Today, the cybersecurity threats they face from outside their organization and the internal needs to simplify the management of their data protection environment have created an opportunity for savvy IT professionals to add value by adopting backup appliances that leverage analytics, microservices, and scalable storage.




Make the Right Choice between Scale-out and Scale-up Backup Appliance Architectures

Companies are always on the lookout for simpler, most cost-effective methods to manage their infrastructure. This explains, in part, the emergence of scale-out architectures over the last few years as a preferred means for implementing backup appliances. It is as scale-out architectures gain momentum that it behooves companies taking a closer look at the benefits and drawbacks of both scale-out and scale-up architectures to make the best choice for their environment.

Backup appliances primarily ship in two architectures: scale-out and scale-up.  A scale-out architecture is comprised of nodes that are logically grouped together using software that the vendor provides. Each node ships with preconfigured amounts of memory, compute, network ports, and storage capacity. The maximum raw capacities of backup appliances from about a few dozen terabytes to nearly twelve petabytes.

In contract, a scale-up architecture places a controller with compute, memory and network ports in front of storage shelves. A storage shelf may be internal or external to the appliance. Each storage shelf holds a fixed number of disk drives.

Backup appliances based on a scale-up architecture usually require lower amounts of storage capacity for an initial deployment. If an organization needs more capacity, it adds more disk drives to these storage shelves, up to some predetermined, fixed, hard upper limit. Backup appliances that use this scale-up architecture range from a few terabytes of maximum raw capacity to over multiple petabytes of maximum raw capacity.

Scale-out Benefits and Limitations

A scale-out architecture, sometimes referred to as a hyper-converged infrastructure (HCI), enables a company to purchase more nodes as it needs them. Each time it acquires another node, it provides more memory, compute, network interfaces, and storage capacity to the existing solutions. This approach addresses enterprise needs to complete increased backup workloads in the same window of time since they have more hardware resources available to them.

This approach also addresses concerns about product upgrades. By placing all nodes in a single configuration, as existing nodes age or run of capacity, new nodes with higher levels of performance and more capacity can be introduced into the scale-out architecture.

Additionally, an organization may account for and depreciate each node individually. While the solution’s software can logically group physically nodes together, there is no requirement to treat all the physical nodes as a single entity. By treating each node as its own physical entity, an organization can depreciate physical over a three to five-year period (or whatever period its accounting rules allow for.) This approach mitigates the need to depreciate newly added appliances in a shorter time frame as is sometimes required when adding capacity to scale-up appliances.

The flexibility of scale-out solutions can potentially create some management overhead. Using a scale-out architecture, an enterprise should verify that as the number of nodes in the scale-out configuration increases, the solution has a means to automatically load balance the workloads and store backup data across all its available nodes. If not, an enterprise may find it spends an increasing amount of time balancing the backup jobs across its available nodes

An enterprise should also verify that all the nodes work together as one collective entity. For instance, an enterprise should verify that the scale-out solution offers “global deduplication”. This feature dedu­plicates data across all the nodes in the system, regardless of on which node the data resides. If it does not offer this feature, the solution will still deduplicate the data but only on each individual node.

Finally, an enterprise should keep its eye on the possibility of “node sprawl” when using these solutions. These solutions make it easy to grow but an enterprise needs to plan for the optimal way to add each node as individual nodes can vary widely in their respective capacity and performance characteristics.

Scale-up Benefits and Limitations

Backup appliances that use a scale-up architecture have their own sets of benefits and limitations. Three features that currently working in their favor include:

  1. Mature
  2. Well-understood
  3. Widely adopted and used

One other broader backup industry trend currently also works in favor of scale-up architectures. More enterprises use snapshots as their primary backup technique. Using these snapshots as the source for the backup frees enterprises to do backups at almost any time of the day. This helps to mitigate the night and weekend performance bottleneck that can occur when forced to do all backups at the same using one of these appliances as the backup target.

A company may encounter the following challenges when working with scale-up appliances:

First, it must size and configure the appliance correctly. This requires an enterprise to have a good understanding of its current and anticipated backup workloads, its total amount of data to backup, and its data retention requirements. Should it overestimate its requirements, it may end up with an appliance oversized for its environment. Should it underestimate its requirements, backup jobs may not complete on time or it may run out of capacity, requiring it to buy another appliance sooner than it anticipated.

Second, all storage capacity sits behind a single controller. This architecture necessitates that the controller be sufficiently sized to meet all current and future backup workloads. Even though the appliance may support the addition of more disk drives, all backup jobs will still need to run through the same controller. Depending on the amount of data and how quickly backup jobs need to complete, this could bottleneck performance and slow backup and recovery jobs.

Make the Right Backup Appliance Choice

In order to make the right choice between these two architectures, the choice may come down to how well you understand your own environment. If a company expects to experience periods of rapid or unexpected data growth, using a scale-out appliance will often be a better approach. In these scenarios, look to appliances from Cohesity, Commvault, ExaGrid, NEC and StorageCraft.

If a company expects more predictable or minimal data growth in its environment, scale-up backup solutions such as the Asigra TrueNAS and Unitrends appliances will likely better match its requirements.




DCIG Introduces Two New Offerings in 2019

DCIG often gets so busy covering all the new and emerging technologies in multiple markets that we can neglect to inform our current and prospective clients of new offerings that DCIG has brought to market. Today I address this oversight.

While many of you know DCIG for its Buyer’s Guides, blogs, and executive white papers, DCIG now offers the following two assets that companies can contract DCIG to create:

1.      DCIG Competitive Intelligence Reports. These reports start by taking a subset of the information we gather as part of creating the DCIG Buyer’s Guides. These reports compare features from two to five selected products and examines how they deliver on these features. The purpose of these reports is not to declare which feature implementation is “best”. Rather, it examines how each product implements these select features and what the most appropriate use case is for those features.

2.      DCIG Content Bundle. In today’s world, people consume the same content in multiple ways. Some prefer to hear it via podcasts. Some prefer to watch it on video. Some want to digest it in bite size chunks in blog entries. Still others want the whole enchilada in the form of a white paper. To meet these various demands, DCIG delivers the same core set of content in all four of these formats as part of its newly created content bundle.

If any of these new offerings pique your interest, let us know! We would love to have the opportunity to explain how they work and provide you with a sample of these offerings. Simply click on this link to send us an email to inquire about these services.




The New Need to Create a Secondary Perimeter to Detect for Malware’s Presence

Malware – and specifically ransomware – tends to regularly make headlines with some business somewhere in the world reporting having its data encrypted by it. Due to this routine occurrence, companies need to acknowledge that their standard first line defenses such as cybersecurity and backup software no longer completely suffice to detect malware. To augment these defenses, companies need to take new steps to shore up these traditional defenses which, for many, will start with creating a secondary perimeter around their backup stores to detect the presence of malware.

The size of companies getting infected by malware are not what one may classify as “small.” By way of example, a story appeared earlier this week about an 800-bed hospital in Malvern, Australia, that had the medical records of 15,000 of its patients in its cardiology unit compromised and encrypted at the end of January 2019.

While I am unfamiliar with both this hospital’s IT staff and procedures and the details of this incident, one can make two educated observations about its IT operations:

  • One, the hospital is sufficiently large that it likely had anti-virus software and firewalls in place that, in a perfect world, would have detected the malware and thwarted it.
  • Two, it probably did regular backups of its production data (nightly or weekly.) Even if the malware attack did succeed, it should have been able to use backups to recover.

So, the questions become:

  1. Why is his hospital, or any company for that matter, still susceptible to something as theoretically preventable as a malware attack in the form of ransomware?
  2. Why could the hospital not use its backups to recover?

Again, sufficient details are not yet publicly available about this attack to know with certainty why these defenses failed or if they were even in place. If one or both these defenses were not in place, then this hospital was susceptible to becoming a victim to this sort of attack. But even if both these defenses were in place or even if just one was in place, it begs asking, “Why did one or both of these defenses not suffice?

The short answer is, both these defenses remain susceptible to malware attacks whether used separately or together. This deficiency does not necessarily originate with poorly designed anti-virus software, backup software or firewalls. Rather, malware’s rapid evolution and maturity challenges the ability of cybersecurity and backup software providers to keep pace with them.

A 2017 study published by G DATA security experts revealed they discovered a new malware strain about every four seconds.  This massive number of malware strains makes it improbable that anti-virus software and firewalls can alone identify every new strain of malware as it enters a company.  Malware’s rapid evolution can also result in variations of documented ransomware strains such as Locky, NotPetya, and WannaCry slipping through undetected.

Backup software is also under attack by malware. Strains of malware now exist that may remain dormant and undetected for some time. Once inside a company, it first infects production files over a period of days, weeks or even months before it detonates. During the malware’s incubation period, companies will back up these infected production files. At the same time, they will, as part of their normal backup operations, delete their expiring backups.

After a few weeks or months of routine backup operations, all backups created during this time will contain infected production files. Then when the malware does detonate in the production environment, companies may get caught in a Zero-day Attack Loop.

Using cybersecurity software on the perimeter of corporate IT infrastructures and backup software inside the IT infrastructure does help companies detect and prevent malware attacks as well as recover from them. However, the latest strains of malware’s reflect its continuing evolution and growing sophistication that better equips them to bypass these existing corporate countermeasures as is evidenced by attacks on this hospital in Australia and the ones too numerous to mention around the world.

For these reasons, backup software that embeds artificial intelligence, machine learning, and, yes, even cybersecurity software, is entering the market place. Using these products, companies can create a secondary defense perimeter inside their company around their data stores that provides another means for companies to detect existing and new strains of malware as well as better position them to successfully recover from malware attacks when they do occur.




Tips to Selecting the Best Cloud Backup Solution

The cloud has gone mainstream with more companies than ever looking to host their production applications with general-purpose cloud providers such as the Google Cloud Platform (GCP). As this occurs, companies must identify backup solutions architected for the cloud that capitalize on the native features of each provider’s cloud offering to best protect their virtual machines (VMs) hosted in the cloud.

Company that move their applications and data to the cloud must orchestrate the protection of their applications and data once they move them there. GCP and other cloud providers offer highly available environments and replicate data between data centers in the same region. They also provide options in their clouds for companies to configure their applications to automatically fail over, fail back, scale up, and scale back down as well as create snapshots of their data.

To fully leverage these cloud features, companies must identify an overarching tool that orchestrates the management of these availability, backup and recovery features as well as integrates with their applications to create application-consistent backups. To select the right cloud backup solution for them, here are a few tips to help companies do so.

Simple to Start and Stop

The cloud gives companies the flexibility and freedom to start and stop services as needed and then only pay for these services as they use them. The backup solution should give companies the same ease to start and stop these services. It should only bill companies for the applications it protects during the time it protects them.

The simplicity of the software’s deployment should also extend to its configuration and ongoing management. Companies can quickly select and deploy the compute, networking, storage, and security services cloud providers offer. In the same way, the software should similarly make it easy for companies to select and configure it for the backup of VMs. They can also optionally turn the software off if needed.

Takes Care of Itself

When companies select any cloud provider’s service, companies get the benefits of the service without the maintenance headaches associated with owning it. For example, when companies choose to host data on GCP’s Cloud Storage service, they do not need to worry about administering Google’s underlying IT infrastructure. The tasks of replacing faulty HDDs, maintaining HDD firmware, keeping its Cloud Storage OS patched, etc. fall to Google.

In the same way, when companies select backup software, they want its benefits without the overhead of patching it, updating it, and managing it long term. The backup software should be available and run as any other cloud service. However, in the background, the backup software provider should take care of its software’s ongoing maintenance and updates.

Integrates with the Cloud Provider’s Identity Management Services

Companies use services such as LDAP or Microsoft AD to control access to corporate IT resources. Cloud providers also have their own identity management services that companies can use to control their employees’ access to cloud resources.

The backup software will ideally integrate with the cloud provider’s native identity management services to simplify its management and ensure that those who administer the backup solution have permission to access VMs and data in the cloud.

Integrates with the Cloud Provider’s Management Console

Companies want to make their IT environments easier to manage. For many, that begins with a single pane of glass to manage their infrastructure. In cloud environments, companies must adhere to this philosophy as cloud providers offer dozens of cloud services that individuals can view and access through that cloud provider’s management console.

To ensure cloud administrators remain aware that the backup is available as an option, much less use it, the backup software must integrate with the cloud provider’s default management console. In this way, these individuals can remember to use it and easily incorporate its management into their overall job responsibilities.

Controls Cloud Costs

It should come as no great surprise that cloud providers make their money when companies use their services. The more of their services that companies use, the more the cloud providers charge. It should also not shock anyone the default services that cloud providers offer may be among their most expensive.

The backup software can help companies avoid racking up unneeded costs in the cloud. The backup software will primarily consume storage capacity in the cloud. The software should offer features that help manage these costs. Aside from having policies in place to tier backup data as its ages across these different storage types, it should also provide options to archive, compress, deduplicate, and even delete data. Ideally, it will also spin up cloud compute resources when needed and shut them down once backup jobs complete to further control costs in the cloud.

HYCU Brings the Benefits of Cloud to Backup

Companies choose the cloud for simple reasons: flexibility, scalability, and simplicity. They already experience these benefits when they choose the cloud’s existing compute, networking, storage, and security services. So, they may rightfully wonder, why should the software service they use to orchestrate their backup experience in the cloud be any different?

In short, it should not be any different. As companies adopt and adapt to the cloud’s consumption model, they will expect all services they consume in the cloud to follow its billing and usage model. Companies should not give backup a pass on this growing requirement.

HYCU is the first backup and recovery solution that companies can choose when protecting applications and data on the Google Cloud Platform to follow these basic principles of consuming cloud services. By integrating with GCP’s identity management services, being simple to start and stop, and helping companies control their costs, among others, HYCU exemplifies how easy backup and recovery can and should be in the cloud. HYCU provides companies with the breadth of backup services that their applications and data hosted in the cloud need while relieving them of the responsibility to continue to manage and maintain it.




Number of Appliances Dedicated to Deduplicating Backup Data Shrinks even as Data Universe Expands

One would think that with the continuing explosion in the amount of data being created every year, the number of appliances that can reduce the amount of data stored by deduplicating it would be increasing. That statement is both true and flawed. On one hand, the number of backup and storage appliances that can deduplicate data has never been higher and continues to increase. On the other hand, the number of vendors that create physical target-based appliances dedicated to the deduplication of backup data continues to shrink.

Data Universe Expands

In November 2018 IDC released a report where it estimated the amount of data that will be created, captured, and replicated will increase five-fold from the current 33 zettabytes (ZBs) to about 175 ZBs in 2025. Whether one agrees with that estimate, there is little doubt that there are more ways than ever in which data gets created. These include:

  • Endpoint devices such as PCs, tablets, and smart phones
  • Edge devices such as sensors that collect data
  • Video and audio recording devices
  • Traditional data centers
  • The creation of data through the backup, replication and copying of this created data
  • The creation of metadata that describes, categorizes, and analyzes this data

All these sources and means of creating data means there is more data than ever under management. But as this occurs, the number of the products originally developed to control this data growth – hardware appliances that specialize in the deduplication of backup data after it is backed up such as those from Dell EMC, ExaGrid, and HPE – has shrunk in recent years.

Here are the top five reasons for this trend.

1. Deduplication has Moved onto Storage Arrays.

Many storage arrays, both primary and secondary, give companies the option to deduplicate data. While these arrays may not achieve the same deduplication ratios as appliances purpose-built for the deduplication of backup data, their combination of lower costs and highs levels of storage capacity offset the inabilities of their deduplication software to optimize backup data.

2. Backup software offers deduplication capabilities.

Rather than waiting to deduplicate backup data on a hardware appliance, almost all enterprise backup software products can deduplicate on either the client or the backup server before storing it. This eliminates the need to use a storage device dedicated to deduplicating data.

3. Virtual appliances that perform deduplication on the rise.

Some providers, such as Quest Software, have exited the physical deduplication backup target appliance market and re-emerged with virtual appliances that deduplicate data. These give companies new flexibility to use hardware from any provider they want and implement their software-defined data center strategy more aggressively.

4. Newly created data may not deduplicate well or at all.

A lot of the new data that companies may not deduplicate well or at all. Audio or video files may not change and will only deduplicate if full backups are done – which may be rare. Encrypted data will not deduplicate at all. In these circumstances, deduplication appliances are rarely if ever needed.

5. Multiple backup copies of the data may not be needed.

Much of the data collected from edge and endpoint devices may only need a couple of copies of data, if that. Audio and video files may also fall into this same category of not needing to retain more than a couple copies of data. To get the full benefits of a target-based deduplication appliance, one needs to backup the same data multiple times – usually at least six times if not more. This reduced need to backup and retain multiple copies of data diminishes the need for these appliances.

Remaining Deduplication Appliances More Finely Tuned for Enterprise Requirements

The reduction in the number of vendors shipping physical target-based deduplication backup appliances seems almost counter-intuitive in the light of the ongoing explosion in data growth that we are witnessing. But when one considers must of data being created and its corresponding data protection and retention requirements, the decrease in the number of target-based deduplication appliances available is understandable.

The upside is that the vendors who do remain and the physical target-based deduplication appliances that they ship are more finely tuned for the needs of today’s enterprises. They are larger, better suited for recovery, have more cloud capabilities, and account for some of these other broader trends mentioned above. These factors and others will be covered in the forthcoming DCIG Buyer’s Guide on Enterprise Deduplication Backup Appliances.




The Early Implications of NVMe/TCP on Ethernet Network Designs

The ratification in November 2018 of the NVMe/TCP standard officially opened the doors for NVMe/TCP to begin to find its way into corporate IT environments. Earlier this week I had the opportunity to listen in on a webinar that SNIA hosted which provided an update on NVMe/TCP’s latest developments and its implications for enterprise IT. Here are four key takeaways from that presentation and how these changes will impact corporate data center Ethernet network designs.

First, NVMe/TCP will accelerate the deployment of NVMe in enterprises.

NVMe is already available in networked storage environments using competing protocols such as RDMA which ships as RoCE (RDMA over Converged Ethernet). The challenge is no one (well, very few anyway) use RDMA in any meaningful way in their environment so using RoCE to run NVMe never gained and will likely never gain any momentum.

The availability of NVMe over TCP changes that. Companies already understand TCP, deploy it everywhere, and know how to scale and run it over their existing Ethernet networks. NVMe/TCP will build on this legacy infrastructure and knowledge.

Second, any latency that NVMe/TCP introduces still pales in comparison to existing storage networking protocols.

Running NVMe over TCP does introduces latency versus using RoCE. However, the latency that TCP introduces is nominal and will likely be measured in microseconds in most circumstances. Most applications will not even detect this level of latency due to the substantial jump in performance that natively running NVMe over TCP will provide versus using existing storage protocols such as iSCSI and FC.

Third, the introduction of NVMe/TCP will require companies implement Ethernet network designs that minimize latency.

Ethernet networks may implement buffering in Ethernet switches to handle periods of peak workloads. Companies will need to modify that network design technique when deploying NVMe/TCP as buffering introduces latency into the network and NVMe is highly latency sensitive. Companies will need to more carefully balance how much buffering they introduce on Ethernet switches.

Fourth, get familiar with the term “incast collapse” on Ethernet networks and how to mitigate it.

NVMe can support up to 64,000 queues. Every queue that NVMe opens up initiates a TCP session. Here is where challenges may eventually surface. Simultaneously opening up multiple queues will result in multiple TCP sessions initiating at the same time. This could, in turn, have all these sessions arrive at a common congestion point in the Ethernet network at the same time. The network remedies this by having all TCP sessions backing off at the same time, or an incast collapse, creating latency in the network.

Source: University of California-Berkeley

Historically this has been a very specialized and rare occurrence in networking due to the low probability that such an event would ever take place. But the introduction of NVMe/TCP into the network makes the possibility of such a event much more likely to occur, especially as more companies deploy NVMe/TCP into their environment.

The Ratification of the NVMe/TCP

Ratification of the NVMe/TCP standard potentially makes every enterprise data center a candidate for storage systems that can deliver dramatically better performance to their work loads. Until the performance demands of every workload in a data center are met instantaneously, some workload requests will queue up behind a bottleneck in the data center infrastructure.

Just as introducing flash memory into enterprise storage systems revealed bottlenecks in storage operating system software and storage protocols, NVMe/TCP-based storage systems will reveal bottlenecks in data center networks. Enterprises seeking to accelerate their applications by implementing NVMe/TCP-based storage systems may discover bottlenecks in their networks that need to be addressed in order to see the full benefits that NVMe/TCP-based storage.

To view this presentation in its entirety, follow this link.




All-inclusive Software Licensing: Best Feature Ever … with Caveats

On the surface, all-inclusive software licensing sounds great. You get all the software features that the product offers at no additional charge. You can use them – or not use them – at your discretion. It simplifies product purchases and ongoing licensing.

But what if you opt not to use all the product’s features or only need a small subset of them? In those circumstances, you need to take a hard look at any product that offers all-inclusive software licensing to determine if it will deliver the value that you expect.

Why We Like All-Inclusive Software Licensing

All-inclusive software licensing has taken off in recent years with more enterprise data storage and data protection products than ever delivering their software licensing in this manner. Further, this trend shows no signs of abating for the following reasons:

  • It makes lives easier for the procurement since they do not have manage and negotiate software licensing separately.
  • It makes lives easier for the IT staff who want to use its features only to find out they cannot use them because they do not have a license to use them.
  • It helps the vendors because their customers use their features. The more they use and like the features, the more apt they are to keep using the product long term.
  • It provides insurance for the companies involved that if they do unexpectedly need a feature, they do not have to go back to the proverbial well and ask for more money to license it.
  • It helps IT be more responsive to changes in business requirements. Business need can change unexpectedly. It happens where IT is assured that a certain feature will never be of interest to the end user. Suddenly, this “never gonna need it” becomes a “gotta have it” requirement.

All-inclusive software licensing solves these dilemmas and others.

The Best Feature Ever … Has Some Caveats

The reasons as to why companies may consider all-inclusive software licensing the best feature ever are largely self-evident. But there are some caveats as to why companies should minimally examine all-inclusive software licensing before they select any product that supports it.

  1. Verify you will use the features offered by the platform. It is great that a storage platform offers deduplication, compression, thin provisioning, snapshots, replication, metro clusters, etc., etc. at no extra charge. But if you do not use these features now and have no plans to use them, guess what? You are still going to indirectly pay for them if you buy the product.
  2. Verify the provider measures and knows which of its features are used. When you buy all-inclusive software licensing, you generally expect the vendor to support it and continue to develop it. But how does the vendor know which of its features are being used, when they are being used, and for what purposes? It makes no sense for the provider to staff its support lines with experts in replication or continue developing its replication features if no one uses it. Be sure you select a product that regularly monitors and reports back to the providers which of its features are used, how they are used and actively supports and develops them.
  3. Match your requirements to the features available on the product. It still pays to do your homework. Know your requirements and then evaluate products with all-inclusive software licensing based upon them.
  4. Verify the software works well in your environment. I have run across a few providers who led the way in providing all-inclusive software licensing. Yet the ones who selected the product based on this offering found out the features were not as robust as they anticipated or were so difficult to use that they had to abandon using them. In short, having a license to use software that does not work in your environment does not help anyone.
  5. Try to quantify if other companies use the specific software features. Ideally, you want to know that others like you use the feature in production. This can help you avoid become an unsuspecting beta-tester for that feature.

Be Grateful but Wary

I, for one, am grateful that providers have come around with more of them making all-inclusive software licensing available as a licensing option for their products. But the software features that vendors include with their all-inclusive software licensing vary from product to product. They also differ in their maturity, robustness, and fullness of support.

It behooves everyone to hop on the all-inclusive software licensing bandwagon. But as you do, verify to which train you hitched your wagon and that it will take you to where you want to go.




Time: The Secret Ingredient behind an Effective AI or ML Product

In 2019 the level of interest that companies expressed in using artificial Intelligence (AI) and machine learning (ML) exploded. Their interest is justifiable. These technologies gather the almost endless streams of data coming out of the scads of devices that companies deploy everywhere, analyze it, and then turn it into useful information. But time is the secret ingredient that companies must look for as they look to select an effective AI or ML product.

Data
Collection Must Proceed AI and ML

The premise behind the deployment of AI and ML technologies
is sound. Every device that a company deploys, in whatever form it takes (video
camera, storage array, server, network switch, automatic door opener, whatever)
has some type of software on it. This software serves two purposes:

  1. Operates the device
  2. Gathers data about the device’s operations, health, and potentially even the environment in which it operates

Option 1 initially drove the development and deployment of
the device’s software while Option 2 sometimes got characterized as a necessary
evil to identify and resolve issues with the device before the device was
impacted. But with more devices Internet enabled, the data each device gathers no
longer needs to remain stranded on each device. It could be centralized.

Devices can now send their data to a central data repository. This is often hosted and supported by the device manufacturer though companies can do this data collection and aggregation on their own.

This is where the AI and ML comes into the picture. Once collected, the manufacturers use AI or ML software to analyze this aggregated amount of data. This analysis can reveal broader trends and patterns otherwise undetectable if the data remained on the devices.

Only
Time Can Deliver an Effective AI or ML Strategy

But here is a key to choosing a product that is truly effective at delivering AI and ML. The value that AI and ML technologies bring relies upon having devices deployed and in production in the field for some time. New vendors, products, and even new deployments, even when they offer AI and ML features, may not provide meaningful insights until the devices collect and analyze a large amount of data over some time from these devices. This can take months or perhaps even years to accomplish.

Only after data is collected and analyzed will the full value of AI or ML technologies become fully evident. Initially, they may help anticipate and prevent some issues. But their effectiveness at anticipating and predicting issues will be limited until they have months or years worth of data at their disposal to analyze.

The evidence of this is seen from companies such as HPE Nimble and Unitrends, among others. Each has improved its ability to better support its clients and resolve issues before companies even know they have issues. For example, HPE Nimble and Unitrends each use their respective technologies to identify and resolve many hardware issues before they impact production.

In each example, each provider needed to collect a great deal of data over multiple years and analyze it before they could proactively and confidently take the appropriate actions to predict and resolve specific issues.

This element of time gives the manufacturers who have large numbers of devices already in the field and who offer AI and ML such a substantial head start in this race to be the leaders in AI and MO. Those just deploying these technologies will still need to gather data for some time period from multiple data points before they can provide the broad type of analytics that companies need and are coming to expect.




HYCU-X Piggybacks on Existing HCI Platforms to Put Itself in the Scale-out Backup Conversation

Vendors are finding multiple ways to enter the scale-out hyper-converged infrastructure (HCI) backup conversation. Some acquire other companies such as StorageCraft did in early 2017 with its acquisition of ExaBlox. Others build their own such as Cohesity and Commvault did. Yet among these many iterations of scale-out, HCI-based backup systems, HYCU’s decision to piggyback its new HYCU-X on top of existing HCI offerings, starting with Nutanix’s AHV HCI Platform, represents one of the better and more insightful ways to deliver backup using a scale-out architecture.

To say that HYCU and Nutanix were inextricably linked before the HYCU-X announcement almost goes without saying. HYCU was the first to market in June 2017 with a backup solution specifically targeted and integrated with the Nutanix AHV HCI Platform. Since then, HYCU has been a leader in providing backup solutions targeted at Nutanix AHV environments.

In coming out with HYCU-X, HYCU addresses an overlooked segment in the HCI backup space. Companies looking for a scale-out secondary storage systems to use as their backup solution typically had to go with a product that was:

  1. New to the backup market
  2. New to the HCI market; or
  3. New to both the backup and HCI markets.

Of these three, a backup provider that fell into either the 2nd or 3rd category where it was or is in any way new to the HCI market is less than ideal. Unfortunately, this is where most backup products fall as the HCI market itself is still relatively new and maturing.

However, this scenario puts these vendors in a tenuous position when it
comes to optimizing their backup product. They must continue to improve and
upgrade their backup solution even as they try to build and maintain an
emerging and evolving HCI platform that supports it. This is not an ideal
situation for most backup providers as it can sap their available resources.

By HYCU initially delivering HYCU-X built on Nutanix’s AHV Platform, it avoids having to create and maintain separate teams to build separate backup and HCI solutions. Rather, HYCU can rely upon Nutanix’s pre-existing and proven AHV HCI Platform and focus on building HYCU-X to optimize Nutanix AHV Platform for use in this role as a scale-out HCI backup platform. In so doing, both HCYU and Nutanix can strive to continue to deliver features and functions that can be delivered in as little as one-click.

Now could companies use Nutanix or other HCI platforms as a scale-out storage target without HYCU-X? Perhaps. But with HYCU-X, companies get the backup engine they need to manage the snapshot and replication features natively found on the HCI platform.

By HYCU starting with Nutanix, companies can leverage the Nutanix AHV HCI Platform as a backup target. They can then use HYCU-X to manage the data once it lands there. Further, companies can then potentially use HYCU-X to backup other applications in their environment.

While some may argue that using Nutanix instead of purpose-built scale-out secondary HCI solutions from other backup providers will cost more, the feedback that HYCU has received from its current and prospective customer base suggests this the opposite is true. Companies find that by time they deploy these other providers’ backup and HCI solutions, their costs could exceed the costs of a Nutanix solution running HYCU-X.

The scale-out backup HCI space continues to gain momentum for good reason. Companies want the ease of management, flexibility, and scalability that these solutions provide along with the promise that they give for them to make disaster recoveries much simpler to adopt and easier to manage over time.

By HYCU piggybacking initially on the Nutanix AHV HCI Platform to deliver a scale-out backup solution, companies get the reliability and stability of one of the largest, established HCI providers and access to a backup solution that runs natively on the Nutanix AHV HCI Platform. That will be a hard combination to beat.




Best Practices for Getting Ready to Go “All-in” on the Cloud

To ensure an application migration to the cloud goes well or that a company should even migrate a specific application to the cloud requires a thorough understanding of each application. This understanding should encompass what resources the application currently uses as well as how it behaves over time. Here is a list of best practices that a company can put in place for its on-premises applications before it moves any of them to the cloud.

  1. Identify all applications running on-premises. A company may assume it knows what applications it has running in its data center environment. However, it is better to be safe than sorry. Take inventory and actively monitor its on-premises environment to establish a baseline. During this time, identify any new virtual or physical machines that come online.
  2. Quantify the resources used by these applications and when and how they use them. This step ensures that a company has a firm handle on the resources each application will need in the cloud, how much of these resources each one will need, and what types of resources it will need. For instance, simply knowing one needs to move a virtual machine (VM) to the cloud is insufficient. A company needs to know how much CPU, memory, and storage each VM needs; when the application runs; its run-time behavior; and, its periods of peak performance to choose the most appropriate VM instance type in the cloud to host it.
  3. Identify which applications will move and which will stay. Test and development applications will generally top the list of applications that a company will move to the cloud first. This approach gives a company the opportunity to become familiar with the cloud, its operations, and billing. Then a company should prioritize production applications starting with the ones that have the lowest level of impact to the business. Business and mission critical applications should be some of the last ones that a company moves. Applications that will stay on-premises are often legacy applications or those that cloud providers do not support.
  4. Map each application to the appropriate VM instance in the cloud. To make the best choice requires that a company knows both their application requirements and the offerings available from the cloud provider. This can take some time to quantify as Amazon Web Services (AWS) offers over 90 different VM instance types on which a company may choose to host an application while Microsoft Azure offers over 150 VM instance types. Further, each of these provider’s VMs may be deployed as an on-demand, reserved, or spot instance that each has access to multiple types of storage. A company may even look to move to serverless compute. To select the most appropriate VM instance type for each application requires that a company know at the outset the capacity and performance requirements of each VM as well as its data protection requirements. This information will ensure a company can select the best VM to host it as well as appropriately configure the VM’s CPU, data protection, memory, and storage settings.
  1. Determine which general-purpose cloud provider to use. Due to the multiple VM instance types each cloud provider offers and the varying costs of each VM instance type, it behooves a company to explore which cloud provider can best deliver the hosting services it needs. This decision may come down to price. Once it maps each of its applications to a cloud provider’s VM instance type, a company should be able to get an estimate of what its monthly cost will be to host its applications in each provider’s cloud.

Companies have good reasons for wanting to go “all-in” on the cloud as part of their overall business and IT strategies. But integral to both these strategies, a company must also have a means to ensure the stability of this new hybrid cloud environment as well as provide assurances that its cloud costs will be managed and controlled over time. By going “all-in” on software such as Quest Software’s Foglight, a company can have confidence that its decision to go “all-in” on the cloud will succeed initially and then continue to pay-off over time.

A recent white paper by DCIG provides more considerations for going all-in on the cloud to succeed both initially and over time. This paper is available to download by following this link to Quest Software’s website.




Rethinking Your Data Deduplication Strategy in the Software-defined Data Center

Data Centers Going Software-defined

There is little dispute tomorrow’s data center will become software-defined for reasons no one entirely anticipated even as recently as a few years ago. While companies have long understood the benefits of virtualizing the infrastructure of their data centers, the complexities and costs of integrating and managing data center hardware far exceeded whatever benefits that virtualization delivered. Now thanks to technologies such as such as the Internet of Things (IoT), machine intelligence, and analytics, among others, companies may pursue software-defined strategies more aggressively.

The introduction of technologies that can monitor, report on, analyze, and increasingly manage and optimize data center hardware frees organizations from performing housekeeping tasks such as:

  • Verifying hardware firmware compatibility with applications and operating systems
  • Troubleshooting hot spots in the infrastructure
  • Identifying and repairing failing hardware components

Automating these tasks does more than change how organizations manage their data center infrastructures. It reshapes how they can think about their entire IT strategy. Rather than adapting their business to match the limitations of the hardware they choose, they can now pursue business objectives where they expect their IT hardware infrastructure to support these business initiatives.

This change in perspective has already led to the availability of software-defined compute, networking, and storage solutions. Further, software-defined applications such as databases, firewalls, and other applications that organizations commonly deploy have also emerged. These virtual appliances enable companies to quickly deploy entire application stacks. While it is premature to say that organizations can immediately virtualize their entire data center infrastructure, the foundation exists for them to do so.

Software-defined Storage Deduplication Targets

As they do, data protection software, like any other application, needs to be part of this software-defined conversation. In this regard, backup software finds itself well-positioned to capitalize on this trend. It can be installed on either physical or virtual machines (VMs) and already ships from many providers as a virtual appliance. But storage software that functions primarily as a deduplication storage target already finds itself being boxed out of the broader software-defined conversation.

Software-defined storage (SDS) deduplication targets exist that have significantly increased in storage capabilities. By the end of 2018, a few of these software-defined virtual appliances scaled to support about 100TB or more of capacity. But organizations must exercise caution when looking to position these available solutions as a cornerstone in a broader software-defined deduplication storage target strategy.

This caution, in many cases, stems less from the technology itself and more from the vendors who provide these SDS deduplication target solutions. In every case, save one, these solutions originate with providers who focus on selling hardware solutions.

Foundation for Software-defined Data Centers Being Laid Today

Companies are putting plans in place right now to build the data center of tomorrow. That data center will be a largely software-defined data center with solutions that span both on-premises and cloud environments. To achieve that end, companies need to select solutions that have a software-designed focus which meet their current needs while positioning them for tomorrow’s requirements.

Most layers in the data center stack, to include compute, networking, storage, and even applications, are already well down the road of transforming from hardware-centric to software-centric offerings. Yet in the face of this momentous shift in corporate data center environments, SDS deduplication target solutions have been slow to adapt.

It is this gap that SDS deduplication products such as Quest QoreStor look to fill. Coming from a company with “software” in its name, Quest comes without the hardware baggage that other SDS providers must balance. More importantly, Quest QoreStor offers a feature-rich set of services that range from deduplication to replication to support for all major cloud, hardware, and backup software platforms that comes from 10 years of experience in delivering deduplication software.

Free to focus solely on delivering a SDDC solution, Quest QoreStor represents the type of SDS deduplication target that does truly meet the needs of today’s enterprise while positioning them to realize the promise of tomorrow’s software-defined data center.

To read more of DCIG’s thoughts about using SDS deduplication targets in the software-defined data center of tomorrow, follow this link.




Key Differentiators between the HPE StoreOnce 5650 and Dell EMC Data Domain 9300 Systems

Companies have introduced a plethora of technologies into their core enterprise infrastructures in recent years that include all-flash arrays, cloud, hyper-converged infrastructures, object-based storage, and snapshots, just to name a few. But as they do, a few constants remain. One is the need to backup and recover all the data they create.

Deduplication appliances remain one of the primary means for companies to store this data for short-term recovery, disaster recoveries, and long-term data retention. To fulfill these various roles, companies often select either the HPE StoreOnce 5650 or the Dell EMC Data Domain 9300. (To obtain a complimentary DCIG report that compares these two products, follow this link.)

Their respective deduplication appliance lines share many features in common. They both perform inline deduplication. They both offer client software to do source-side deduplication that reduces data sent over the network and improves backup throughput rates. They both provide companies with the option to backup data over NAS or SAN interfaces.

Despite these similarities, key areas of differentiation between these two product lines remain which include the following:

  1. Cloud support. Every company either has or anticipates using a hybrid cloud configuration as part of its production operations. These two product lines differ in their levels of cloud support.
  2. Deduplication technology. Data Domain was arguably the first to popularize widespread use of deduplication for backup. Since then, others such as the HPE StoreOnce 5650 have come on the scene that compete head-to-head with Data Domain appliances.
  3. Breadth of application integration. Software plug-ins that work with applications and understand their data formats prior to deduplicating the data provide tremendous benefits as they improve data reduction rates and decrease the amount of data sent over the network during backups. The software that accompanies the appliances from these two providers has varying degrees of integration with leading enterprise applications.
  4. Licensing. The usefulness of any product hinges on the features it offers, their viability, and which ones are available to use. Clear distinctions between the HPE StoreOnce and Dell EMC Data Domain solutions exist in this area.
  5. Replication. Copying data off-site for disaster recovery and long-term data retention is paramount in comprehensive enterprise disaster recovery strategies. Products from each of these providers offer this but they differ in the number of features they offer.
  6. Virtual appliance. As more companies adopt software-defined data center strategies, virtual appliances have increased appeal.

In the latest DCIG Pocket Analyst Report, DCIG compares the HPE StoreOnce 5650 and Dell EMC Data Domain 9300 product lines and examines how each well these two products fare in their support of these six areas in which DCIG looks at nearly 100 features to draw its conclusions. This report is currently available at no charge for a limited time on DCIG’s partner website, TechTrove. To receive complimentary access to this report, complete a registration form that you can find at this link.




NVMe: Four Key Trends Set to Drive Its Adoption in 2019 and Beyond

Storage vendors hype NVMe for good reason. It enables all-flash arrays (AFAs) to fully deliver on flash’s performance characteristics. Already NVMe serves as an interconnect between AFA controllers and their back end solid state drives (SSDs) to help these AFAs unlock more of the performance that flash offers. However, the real performance benefits that NVMe can deliver will be unlocked as a result of four key trends set to converge in the 2019/2020 time period. Combined, these will open the doors for many more companies to experience the full breadth of performance benefits that NVMe provides for a much wider swath of applications running in their environment.

Many individuals have heard about the performance benefits of NVMe. Using it, companies can reduce latency with response times measured in few hundred microseconds or less. Further, applications can leverage the many more channels that NVMe has to offer to drive throughput to hundreds of GBs per second and achieve millions of IOPs. These types of performance characteristics have many companies eagerly anticipating NVMe’s widespread availability.

To date, however, few companies have experienced the full breadth of performance characteristics that NVMe offers. This stems from:

  • The lack of AFAs on the market that fully support NVMe (about 20%).
  • The relatively small performance improvements that NVMe offers over existing SAS-attached solid-state drives (SSDs); and,
  • The high level of difficulty and cost associated with deploying NVMe in existing data centers.

This is poised to change in the next 12-24 months with four key trends poised to converge that will open up NVMe to a much wider audience.

  1. Large storage vendors getting ready to enter the NVMe market. AFA providers such as Tegile (Western Digital), iXsystems, Huawei, Lenovo, and others ship products that support NVMe. These vendors represent the leading edge of where NVMe innovation has occurred. However, their share of the storage market remains relatively small compared to providers such as Dell EMC, HPE, IBM, and NetApp. As these large storage providers enter the market with AFAs that support NVMe, expect market acceptance and adoption of NVMe to take off.
  2. The availability of native NVMe drivers on all major operating systems. The only two major enterprise operating systems that have currently native NVMe drivers for their OSes are Linux and VMware. However, until Microsoft and, to a lesser degree, Solaris, offer native NVMe drives, many companies will have to hold off on deploying NVMe in their environments. The good news is that all these major OS providers are actively working on NVMe drivers. Further, expect that the availability of these drivers will closely coincide with the availability of NVMe AFAs from the major storage providers and the release of the NVMe-oF TCP standard.
  3. NVMe-oF TCP protocol standard set to be finalized yet in 2018. Connecting the AFA controller to its backend SSDs via NVMe is only one half – and much easier part – of solving the performance problem. The much larger and more difficult problem is easily connecting hosts to AFAs over existing storage networks as it is currently difficult to setup and scale NVMe-oF. The establishment of the NVMe-oF TCP standard will remedy this and facilitate the introduction and use of NVMe-oF between hosts and AFAs using TCP/IP over existing Ethernet storage networks.
  4. The general availability of NVMe-oF TCP offload cards. To realize the full performance benefits of NVMe-oF using TCP, companies are advised to use NVMe-oF TCP offload cards. Using standard Ethernet cards with no offload engine, companies will still see high throughput but very high CPU utilization (up to 50 percent.) Using the forthcoming NVMe-oF TCP offload cards, performance increases by anywhere from 33 to 150 percent versus native TCP cards while only introducing nominal amounts of latency (single to double digit microseconds.)

The business need for NVMe technology is real. While today’s all-flash arrays have tremendously accelerated application performance, NVMe stands poised to unleash another round of up to 10x or more performance improvements. But to do that, a mix of technologies, standards, and programming changes to existing operating systems must converge for mass adoption in enterprises to occur. This combination of events seems poised to happen in the next 12-24 months.




20 Years in the Making, the Future of Data Management Has Arrived

Mention data management to almost any seasoned IT professional and they will almost immediately greet the term with skepticism. While organizations have found they can manage their data within certain limits, when they remove those boundaries and attempt to do so at scale, those initiatives have historically fallen far short if not outright failed. It is time for that perception to change. 20 years in the making, Commvault Activate puts organizations in a position to finally manage their data at scale.

Those who work in IT are loath to say any feat in technology is impossible. If one looks at the capabilities of any handheld device, one can understand why they have this belief. People can pinpoint exactly where they are almost anywhere in the world to within a few feet. They can take videos, pictures, check the status of their infrastructure, text, … you name it, handheld devices can do it.

By way of example, as I write this, I was present to watch YY Lee, SVP and Chief Strategy Officer of Anaplan, onstage at Commvault GO. She explained how systems using artificial intelligence (AI) were able within a very short time, sometimes days, became experts at playing games such as Texas Hold’em and beat the best players in the world at them.

Despite advances such as these in technology, data management continues to bedevil large and small organizations alike. Sure, organizations may have some level of data management in place for certain applications (think email, file servers, or databases,) but when it comes to identifying and leveraging a tool to deploy data management across an enterprise at scale, that tool has, to date, eluded organizations. This often includes the technology firms that are responsible for producing so much of the hardware that stores this data and software that produces it.

The end for this vexing enterprise challenge finally came into view with Commvault’s announcement of Activate. What makes Activate different from other products that promise to provide data management at scale is that Commvault began development on this product 20 years ago in 1998.

During that time, Commvault became proficient in:

  • Archiving
  • Backup
  • Replication
  • Snapshots
  • Indexing data
  • Supporting multiple different operating systems and file systems
  • Gathering and managing metadata

Perhaps most importantly, it established relationships and gained a foothold in enterprise organizations around the globe. This alone is what differentiates it from almost every other provider of data management software. Commvault has 20+ years of visibility into the behavior and requirements of protecting, moving, and migrating data in enterprise organizations. This insight becomes invaluable when viewed in the context of enterprise data management which has been Commvault’s end game since its inception.

Activate builds on Commvault’s 20 years of product development with Activate’s main differentiator being its ability to stand alone apart from other Commvault software. In other words, companies do not first have to deploy Commvault’s Complete Backup and Recovery or any of its other software to utilize Activate.

They can deploy Activate regardless of whatever other backup, replication, snapshot, etc. software product you may have. But because Activate draws from the same code base as the rest of Commvault’s software, companies can deploy it with a great deal of confidence because of the stability of Commvault’s existing code base.

Once deployed, Activate scans and indexes the data across the company’s environment which can include its archives, backups, file servers, and/or data stored in the cloud. Once indexed, companies can do an assessment of the data in their environment in anticipation of taking next steps such as eDiscovery preparation, remediate data privacy risks, and index and analyze data based upon your own criteria.

Today more so than ever companies recognize they need to manage their data across the entirety of their enterprise. Delivering on this requirement requires a tool appropriately equipped and sufficiently mature to meet enterprise requirements. Commvault Activate answers this call as a software product that has been 20 years in the making to provide enterprises with the foundation they need to manage their data going forward.




Purpose-built Backup Cloud Service Providers: A Practical Starting Point for Cloud Backup and DR

The shift is on toward using cloud service providers for an increasing number of production IT functions with backup and DR often at the top of the list of the tasks that companies first want to deploy in the cloud. But as IT staff seeks to “Check the box” that they can comply with corporate directives to have a cloud solution in place for backup and DR, they also need to simultaneously check the “Simplicity,” “Cost-savings,” and “It Works” boxes.

Benefits of Using a Purpose-built Backup Cloud Service Provider

Cloud service providers purpose-built for backup and DR put companies in the best position to check all those boxes for this initial cloud use case. These providers solve their clients’ immediate challenges of easily and cost-effectively moving backup data off-site and then retaining it long term.

Equally important, addressing the off-site backup challenge in this way positions companies to do data recovery with the purpose-built cloud service provider, a general purpose cloud service provider, or on-premises. It also sets the stage for companies to regularly test their disaster recovery capabilities so they can perform them when necessary.

Choosing a Purpose-built Backup Cloud Service Provider

To choose the right cloud service provider for corporate backup and DR requirements companies want to be cost conscious. But they also want to experience success and not put their corporate data or their broader cloud strategy at risk. A purpose-built cloud service provider such as Unitrends and its Forever Cloud solution frees companies to aggressively and confidently move ahead with a cloud deployment for its backup and DR needs.

picture showing date and time of webinar

Join Our Webinar for a Deeper Look at the Benefits of Using a Purpose-built Backup Cloud Service Provider

Join me next Wednesday, October 17, 2018, at 2 pm EST, for a webinar where I take a deeper look at both purpose-built and general purpose cloud service providers. In this webinar, I examine why service providers purpose-built for cloud backup and DR can save companies both money and time while dramatically improving the odds they can succeed in their cloud backup and DR initiatives. You can register by following this link.

 




HPE Expands Its Big Tent for Enterprise Data Protection

When it comes to the mix of data protection challenges that exist within enterprises today, these companies would love to identify a single product that they can deploy to solve all their challenges. I hate to be the bearer of bad news, but that single product solution does not yet exist. That said, enterprises will find a steadily improving ecosystem of products that increasingly work well together to address this challenge with HPE being at the forefront of putting up a big tent that brings these products together and delivers them as a single solution.

Having largely solved their backup problems at scale, enterprises have new freedom to analyze and address their broader enterprise data protection challenges. As they look to bring long term data retention, data archiving, and multiple types of recovery (single applications, site fail overs, disaster recoveries, and others) under one big tent for data protection, they find they often need to deploy multiple products.

This creates a situation where each product addresses specific pain points that enterprises have. However, multiple products equate to multiple management interfaces that each have their own administrative policies with minimal or no integration between them. This creates a thornier problem – enterprises are left to manage and coordinate the hand-off of the protection and recovery of data between these different individual data protection products.

A few years HPE started to build a “big tent” to tackle these enterprise data protection and recovery issues. It laid the foundation with its HPE 3PAR StoreServ storage arrays, StoreOnce deduplication storage systems, and Recovery Manager Central (RMC) software to help companies coordinate and centrally manage:

  • Snapshots on 3PAR StoreServ arrays
  • Replication between 3PAR StoreServ arrays
  • The efficient movement of data between 3PAR and StoreOnce systems for backup, long term retention, and fast recoveries

This week HPE expanded its big tent of data protection to give companies more flexibility to protect and recover their data more broadly across their enterprise. It did so in the following ways:

  • HPC RMC 6.0 can directly recover data to HPE Nimble storage arrays. Recoveries from backups can be a multi-step process that may require data to pass through the backup software and the application server before it lands on the target storage array. Beginning December 2018, companies can use RMC to directly recover data to HPE Nimble storage arrays from an HPE StoreOnce system without going through the traditional recovery process just as they can already do to HPE 3PAR StoreServ storage arrays.
  • HPE StoreOnce can directly send and retrieve deduplicated data from multiple cloud providers. Companies sometimes fail to consider that general purpose cloud service providers such as Amazon Web Services (AWS) or Microsoft Azure make no provisions to optimize data stored with them such as deduplicating it. Using HP StoreOnce’s new direct support for AWS, Azure, and Scality, companies can use StoreOnce to first compress and deduplicate data before they store the data in the cloud.
  • Integration between Commvault and HPE StoreOnce systems. Out of the gate, companies can use Commvault to manage StoreOnce operations such as replicating data between StoreOnce systems as well as moving data directly from StoreOnce systems to the cloud. Moreover, as this relationship between Commvault and HPE matures, companies will also be able to use HPE’s StoreOnce Catalyst, HPE’s client-based deduplication software agent, in conjunction with Commvault to backup data on server clients where data may not reside on HPE 3PAR or Nimble storage. Using the HPE StoreOnce Catalyst software, Commvault will deduplicate data on the source before sending it to an HPE StoreOnce system.

Source:HPE

Of these three announcements that HPE made this week, this new relationship with Commvault that accompanies its pre-existing relationships with Micro Focus (formerly HPE Data Protector) and Veritas demonstrate HPE’s commitment to helping enterprises build a big tent for their data protection and recovery initiatives. Storing data on the HPE 3PAR and Nimble and using RMC to manage their backups and recoveries on the StoreOnce systems certainly accelerates and simplifies these functions when companies can do so. But by working with these other partners, it illustrates that HPE recognizes that companies will not store all their data on its systems and that HPE will accommodate companies so they can create a single, larger data protection and recovery solution for their enterprise.




Two Hot Technologies to Consider for Your 2019 Budgets

Hard to believe but the first day of autumn is just two days away and with fall weather always comes cooler temperatures (which I happen to enjoy!) This means people are staying inside a little more and doing those fun, end of year activities that everyone enjoys – such as planning their 2019 budgets. As you do so, solutions from BackupAssist and StorMagic are two hot new technologies for companies to consider making room for in the New Year.

BackupAssist 365.

BackupAssist 365 backs up files and emails stored in the cloud. While backup of cloud-based data may seem rather ho-hum in today’s artificial intelligent, block chain obsessed, digital transformation focused world, it solves a real world that nearly every size organization faces: how to cost-effectively and simply protect all those pesky files and emails that people store in cloud applications such as DropBox, Office 365, Google Drive, OneDrive, Gmail, Outlook and others.

To do so, BackupAssist 365 adopted two innovative yet practical approaches to protect files and emails.

  • First, it interfaces directly with these various cloud providers to backup this data. Using your login permissions (which you provide when configuring the software,) BackupAssist 365 accesses data directly in the cloud. This negates the need for your server, PC, or laptop to be turned on when these backups occur so backups can occur at any time.
  • Second, it does cloud-to-local In other words, rather than running up more data transfer and network costs that come with backing up to another cloud, it backs the data backup to local storage on your site. While that may seem a little odd in today’s cloud-centric world, companies can get a great deal of storage capacity for nominal amounts of money. Since it only does an initial full backup and then differential backups thereafter, the ongoing data transfer costs are nominal and the amount of storage capacity that one should need onsite equally small.

Perhaps the best part about BackupAssist 365 is its cost (or lack thereof.) BackupAssist 365 licenses its software on a per user basis with each user email account counting as one user license. However, this one email account covers the backup of that user’s data in any cloud service used by that user. Further, the cost is only $1/month per user with a decreasing cost for greater number of users. In fact, the cost is so low on a per user basis, companies may not even need to budget for this service. They can just start using it and expense their credit cards to keep it below corporate radar screens.

StorMagic SvSAN

The StorMagic SvSAN touches on another two hot technology trends that I purposefully (or not so purposefully) left out above: hyperconverged infrastructure or HCI and edge computing. However, unlike many of the HCI and edge computing plays in the marketplace such as Cisco HyperFlex, Dell EMC VxRail, and Nutanix, StorMagic has not forgotten about cost constraints that branch, remote, and small offices face.

As Cisco, Dell EMC, Nutanix and others chase the large enterprise data center opportunities, they often leave remote, branch, and small offices with two choices: pay up or find another solution. Many of these size offices are opting to find alternative solutions.

This is where StorMagic primarily plays. For a less well-known player, they play much bigger than they may first appear. Through partnerships with large providers such as Cisco and Lenovo among others, StorMagic comes to market with highly available, two-server systems that scale across dozens, hundreds, or even thousands of remote sites. To get a sense of StorMagic’s scalability, walk into any of the 2,000+ Home Depots in the United States or Mexico and ask to look at the computer system that hosts their compute and storage. If the Home Depot lets you and you can find it, you will find a StorMagic system running somewhere in the store.

The other big challenge that each StorMagic system also addresses is security. Because their systems can be deployed almost anywhere in any environment, it does make them susceptible to theft. In fact, in talking to one of its representatives, he shared a story where someone drove a forklift through the side of a building and stole a computer system at one of its customer sites. Not that it mattered. To counter these types of threats, StorMagic encrypts all the data on its HCI solutions with its own software that is FIPS 140-2 compliant.

Best of all, to get these capabilities, companies do not have to break the bank to acquire one of these systems. The list price for the Standard Edition of the SvSAN software, which includes 2TB of usable storage, high availability, and remote management, is $2,500.

As companies look ahead and plan their 2019 budgets, they need to take care of their operational requirements but they may also want to dip their toes in the water to get the latest and greatest technologies. These two technologies give companies the opportunities to do both. Using BackupAssist 365, companies can quickly and easily address their pesky cloud file and email backup challenges while StorMagic gives them the opportunity to affordably and safely explore the HCI and edge computing waters.




Analytics, Automation and Hybrid Clouds among the Key Takeaways from VMworld 2018

At early VMworld shows, stories emerged of attendees scurrying from booth to booth on the exhibit floor looking for VM data protection and hardware solutions to address the early challenges that VMware ESXi presented. Fast forward to the 2018 VMworld show and the motivation behind attendees attending training sessions and visiting vendor booths has changed significantly. Now they want solutions that bring together their private and public clouds, offer better ways to analyze and automate their virtualized environments, and deliver demonstrable cost savings and/or revenue opportunities after deploying them.

The entrance to the VMworld 2018 exhibit hall greeted attendees a little differently this than in year’s past. Granted, there were still some of the usual suspects such as Dell EMC and HPE that have reserved booths at this show for many years. But right alongside them were relative newcomers (to the VMworld show anyway) such as Amazon Web Services and OVHcloud.

Then as one traversed the exhibit hall floor and visited the booths of the vendors immediately behind them, the data protection and hardware themes of the early VMworld shows persisted in these booths, though the messaging and many of the vendor names have changed since the early days of this show.

Companies such as Cohesity, Druva, and Rubrik represent the next generation of data protection solutions for vSphere while companies such as Intel and Trend Micro have a more pronounced presence on the VMworld show floor. Together these exhibitors reflect the changing dynamics of what is occurring in today’s data centers and what the current generation of organizations are looking for vendors to provide for their increasingly virtualized environments. Consider:

  1. Private and public cloud are coming together to become hybrid. The theme of hybrid clouds with applications that can span both public and private clouds began with VMworld’s opening keynote announcing the availability of Amazon Relational Database Service (Amazon RDS) on VMware. Available in the coming months, this functionality will free organizations to automate the setup of Microsoft SQL Server, Oracle, PostgreSQL, MariaDB and MySQL databases in their traditional VMware environments and then migrate them to the AWS cloud. Those interested in trying out this new service can register here for a preview.
  2. Analytics will pave the ways for increasing levels of automation. As organizations of all sizes adopt hybrid environments, the only way they can effectively manage their hybrid environments at scale is to automate their management. This begins with the use of analytics tools that capture the data points coming in from the underlying hardware, the operating systems, the applications, the public clouds to which they attach, the databases, the devices which feed them the data, whatever.

Evidence of growing presence of these analytics tools that enable this automation was everywhere at VMworld. One good example is Runecast analyzes the logs of these environments and then also scours blogs, white papers, forums, and other online sources for best practices to advise companies on how to best configure their environments. Another one is Login VSI which does performance benchmarking and forecasting to anticipate how VDI patches and upgrades will impact the current infrastructure.

  1. The cost savings and revenue opportunities for these hybrid environments promise to be staggering. One of the more compelling segments in one of the keynotes was the savings that many companies initially achieved deploying vSphere. Below is one graphic that appeared at the 8:23 mark in this video of the second day’s keynote where a company reduced its spend on utility charges by over $60,000 per month or an 84% reduction in cost. Granted, this example was for illustration purposes but it seemed inline with other stories I have anecdotally heard.

Source: VMware

But as companies move into this hybrid world that combines private and public clouds, the value proposition changes. While companies may still see cost savings going forward, it is more likely that they will realize and achieve new opportunities that were simply not possible before. For instance, they may deliver automated disaster recoveries and high availability for many more or all their applications. Alternatively, they will be able to bring new products and services to market much more quickly or perform analysis that simply could not have been done before because they have access to resources that were unavailable to them in a cost-effective or timely manner.