Anyone who managed IT infrastructures in the late 1990′s or early 2000′s probably still remembers how external storage arrays were largely a novelty reserved for high end enterprises with big data centers and deep pockets. Fast forward to today and a plethora of storage arrays exist in a variety of shapes and sizes at increasingly low price points. As such it can be difficult to distinguish between them. To help organizations sort them out, my blog entry today provides a primer on the types of storage arrays currently available on the market.
Archiving or backing up large amounts of data to the cloud is very appealing until one starts to examine the mechanics of actually sending or retrieving that data from the cloud. Waiting minutes or hours to send or retrieve data is no longer acceptable to today’s end-users who are rapidly becoming accustomed to near-instant response times. In this fifth and final part of my interview series with BridgeSTOR’s CEO John Matze, he explains how sending or retrieving data in a piecemeal fashion to the cloud is the fastest and most effective way to do so.
Storing archival and backup data in the cloud is high on the list of priorities of many organizations if for no other reason is that the data remains accessible and available without organizations having to bear the burden of managing the data locally long term. But as more organizations use cloud storage gateways to store this data, they will find distinct differences in how these appliances manage data in the cloud with differences sometimes existing even between appliances from the same vendor. In this fourth part of my interview series with BridgeSTOR’s CEO John Matze, he reveals the various methods that the BridgeSTOR NAS and VTL cloud gateway appliances store, access and manage this data locally and in the cloud.
Ask any large organization how many tapes they have sitting around in local storage or at Iron Mountain or some other third party storage facility and odds are they have more tapes – and are likely spending more money storing these tapes – than they would like to admit. This opens up a unique opportunity for a third party provider to solve this dilemma. In this third part of my interview series with BridgeSTOR’s CEO John Matze, we discuss how using the BridgeSTOR VTL cloud gateway appliance organizations can move their tape museums into the cloud.
NAS gateway appliances that connect to backend public storage clouds are still not a “dime a dozen” but they are definitely more prevalent than they were even a few years ago. However a new class of gateway appliances that provides a virtual tape library (VTL) is now available from BridgeSTOR. In this second part of my interview series with BridgeSTOR’s CEO John Matze, we discuss the inner workings of its VTL interface that it is making available this month on its cloud gateway appliance.
Storing data in the cloud is on almost every company’s radar screen. However what remains a little hazy is the best way for them to store data in the cloud. The use of public cloud storage gateways as a means to store and retrieve data from public storage clouds is rapidly emerging as the preferred option to do so. In this first blog entry of my interview series with BridgeSTOR’s CEO John Matze, we take a look at some of the different gateways solutions available for accessing public storage clouds and how they differ.
Security, security, security. The word rings loud and clear in the ears of corporate business owners when it comes to the data and information stored on mobile devices. Allowing employees to bring their own devices (BYOD) may increase productivity but an organization needs to make sure their sensitive information does not become public. In the fourth and final part of my interview with Amtel, Inc. CEO, Pankaj Gupta, we examine the growing trend in enterprises adopting BYOD and the scope of systems management being offered by organizations like Amtel.
On my flight home last night from San Francisco to Omaha I had the opportunity to reflect a bit on what I learned and saw at VMworld 2013 this past week. In particular, I spent a fair amount of time thinking back as to what were the most transformational technologies I saw while at VMworld. Having now thought about it, I have to classify those offered by Amazon, Avere and Oracle as having the most promise to transform how the data center is delivered and designed sooner rather than later.
In the arena of mobile data storage a business has a few options. In choosing which solution is best a company may ask itself, “Is keeping data on premises or putting it in the cloud more cost-effective?” and “Do we feel comfortable using the cloud for our storage needs?” During my first installment with Amtel, Inc. CEO, Pankaj Gupta, we chatted about how companies can adopt an open attitude in the corporate world of bring your own device (BYOD). Now, in my second segment with Amtel’s CEO, we discuss the benefits of storing data in the cloud, how secure cloud storage really is, and how accepting midmarket companies and larger enterprises are of putting their data in the cloud.
There is a tendency among technology providers to sometimes pooh-pooh the virtualization needs of small and midsized businesses and only focus on the needs of the “really big enterprises.” However when one considers that the 900,000+ companies with 20-500 employees in Canada, the UK and US are less than 30% virtualized, a tremendous opportunity exists for the right technology provider to meet their specific needs.
Just because a backup appliance can back up and recover data to the cloud does not mean they all do so equally well. Further complicating the decision process, some companies back up to their own private cloud while others opt to back up and recover from public clouds. In this second part of my interview series with STORServer’s Jarrett Potts, we examine how backup to public and private clouds differ and what features a backup appliance needs to offer to meet these different requirements.
Backup and the cloud are becoming increasingly linked as they solve two issues that have plagued organizations for years: automating the movement of data offsite and providing a cost-effective means to store it there. But just because one can back up to the cloud does not mean all solutions do so equally well. In this first part of an interview series with STORServer’s Jarrett Potts, we examine what specific features a solution needs to offer to effectively back up to the cloud as well as how the solution needs to be constructed.
Few if any in IT dispute the reality of the cloud or the impact that it is having on businesses. But too many in IT are still prone to say “No” to cloud deployments in THEIR business out of fear of the unknown associated with the cloud. This mindset is now becoming more dangerous as it may turn into a self-fulfilling form of digital Darwinism for you and/or your business.
Companies have spent billions if not tens of billions of dollars putting in place the necessary hardware and software to virtualize their data centers. These numbers fail to take into account the countless man hours that they have also spent planning, configuring and implementing these environments. Yet just as many are ready to take a deep breath and enjoy the fruits of their labors, they are coming to the realization that virtualizing their environment and creating private or public clouds only laid the foundation for the real goal of data center automation.
As companies of all sizes move from physical environments to ones that are more highly virtualized (or even entirely virtualized,) everything changes. While “how backups are done” is sometimes viewed as the biggest change, monitoring the virtual environment and leveraging the cloud are becoming higher priorities for end users. In this fourth and final blog entry in my interview series with PHD Virtual’s CMO Steve Kahan, he discusses how virtualization monitoring and the cloud are impacting the future of backup in general and PHD Virtual specifically.
This past week nearly 120 million people cast their vote in the US presidential election. But another election that is privately ongoing in many enterprises right now is how they are going to vote with their pocket book in terms of which private storage cloud they are going to deploy. This week after attending the HDS Influencer Summit in Santa Clara, CA, and seeing what HDS has to offer, HDS has for now edged out its competitors and earned my vote for having the best private storage cloud for enterprises.
The role of the Chief Information Officer (CIO) is evolving as enterprises worldwide attempt to navigate their way through the fundamental changes required to keep pace with the explosion of Cloud Computing, Social Media, Big Data and Mobile Computing. Information Governance, Compliance, eDiscovery, Data Security and Business Intelligence are now more important than ever. If the CIO can’t keep pace, the fate of the entire enterprise may be at stake.
In this final installment of our interview series with GroupLogic we look at how mobilEcho enables organizations to wipe only company data off employees’ mobile devices, leaving the employee’s personal data untouched. We also hear how GroupLogic is helping telecommunications firms increase revenue by adding value to their customers’ telecommunications experience, and we gain insight into the DNA of the company as a software provider for over 20 years.
It is no longer a matter a question of “if” most organizations are going to implement a private cloud; it is more a matter of “when” and “how to best proceed.” This is where it can get a little hazy as it is not always clear what path an organization should follow to ensure it ends up with a private cloud that meets its needs. While this path is not the same for every organization, there are three principles that organizations may follow to have a high degree of assurance that they will end up with a private cloud that meets their needs.
Many businesses either rely on in-house custom software or purchase and customize commercial software packages to meet their mission critical business needs. Unfortunately these approaches are failing to meet the needs of enterprises as they demand faster development and deployment times for these apps. To address that challenge, a new group of vendors in the cloud promises to improve these mediocre mission critical business processes by speeding up development/deployment cycles while reducing the total cost of ownership of enterprise applications.