The driver that wins the big race does not get as much credit for the win if he simply drove the fastest car on the track. After all, logic says the driver should win. A recent trend in data storage has been to migrate to all flash array storage solutions. Flash drives are certainly faster than HDD disks. For those enterprises that implement AFA storage arrays, performance is definitely fast, which it should be using the logic of the fastest racecar. However, there are companies that are achieving ultrafast performance without having to pay the premium price for high performance storage media. Now that is real innovation.
Posts by Jay Cardin
In order to maintain a competitive edge, businesses need to be able to move fast, and recover faster. When your success depends on evaluating data in real-time, any kind of hiccup can cost you money. Today’s enterprises need more than backup – they need to be able to rapidly restore critical data in the right contexts. That’s why more and more frequently, disk-to-disk-to-tape (D2D2T) architectures are giving way to faster, more efficient flash-to-flash-to-cloud (F2F2C) technologies.
Enterprises are increasingly realizing the value of their data, and are capturing as much of it as possible to glean whatever insights they can. The challenge, however, is that IT budgets are not growing at the rates needed to properly handle these growing mountains of digital gold.
In order for developers to keep up with the speed of business, new innovations in storage, backup, and disaster recovery are being rolled out at a breakneck pace. To stay competitive, enterprises must adopt these new methods and technologies for faster restores and greater opportunities to leverage the true value of data across their organization.
Digital transformation is reinventing the IT infrastructure. Enterprises are shifting to software-defined data centers and the new technology that comes with them. AI and machine learning (ML), cloud utilization, and software-defined storage are just a few of the many new tools that organizations now employ to help them harness the power of their data. As a result, IT infrastructure management has become increasingly complex. IT staff can no longer be expected to manage every part of the infrastructure. However, the answer to this issue is simple: let technology manage itself. AI/ML intelligent data platforms can provide comprehensive data management throughout the infrastructure.
Digital transformation is fundamentally changing the way organizations view and manage their data, with an estimated 60 percent of data-driven enterprises scheduled to begin digital transformation by 2020. This transformation allows enterprises to use their data to generate revenue and keep a competitive edge – but only if it can be stored accurately and organized in a way that keeps it easily accessible.
From news headlines to television sitcom story lines, ransomware has become a major player in the world of IT security. High-profile attacks against enterprises are on the rise, their numbers dramatically increasing every year with nearly two-thirds of organizations surveyed reporting an attack in the last year, and 22 percent reporting weekly attacks.
Across industry lines, there’s a trend of enterprises moving their applications to the public cloud. The cost of cloud computing and storage have dropped significantly over the past few years, and IT leaders are seeing the value of not owning physical infrastructure.
The old fable of the lion and the gazelle is a great analogy for business today. Each day the lion must outrun the slowest gazelle or starve. Each day the gazelle must outrun the fastest lion or perish. Whether you are a lion or a gazelle, when the sun comes up, you best be running. The same is true for your business because tomorrow, you best be running. The main difference for many businesses today is it’s not just for when the sun comes up. It is all the time.
How resilient is your organization from an IT perspective? The Merriam-Webster Dictionary defines resilience as, “an ability to recover from or adjust easily to misfortune or change.” With the digital transformations of so many companies today, resiliency to change and disruption needs to be a priority. What does resilience mean for the enterprise today? It means having the ability to seamlessly adapt to change while enduring and responding to unplanned events such as:
In order to prevail in the globally competitive economy today, companies are continually examining their processes, operations, and infrastructure to find areas which add little or no value to the business. For enterprises today, one of those areas is data protection. Think about your backups. Yes, they can potentially “save the day” in the event of a failed server, ransomware attack, or natural disaster. The problem of course is that your backups just sits there idle, indolently waiting for a bad day to occur. Your backups are part of an expensive insurance policy that consumes a lot of resources.
Software as a service and many other digital business models have never been the same since the possibility of leveraging hybrid IT. Thanks to it, an organization can deliver services in a more optimized, balanced, automated, granular, and flexible fashion.
IDC completed a web survey involving 83 end users that have purchased and deployed Hewlett Packard Enterprise (HPE) SimpliVity solutions in their organizations. IDC also gathered qualitative data by conducting in-depth phone interviews with three customers using SimpliVity in production environments. This post aims to share the results of those surveys to show you what customers are saying first hand.
New buzz words tend to come and go, especially as enterprise storage vendors promote their offerings to prospective customers. A few years’ back, we started to hear a lot about scale-out storage vs. scale-up storage. The popularity of scale-out storage has since grown, as have the questions about the difference between scale-up, scale-out, and related terms like hyperconverged and converged infrastructures.
The face of data storage in enterprise data centers has changed in the past few years with the rise to prominence of solid-state, or flash, storage. This advancement of storage technology has now become so widespread among enterprise IT infrastructures around the world that 49% of organizations surveyed by the Enterprise Strategy Group indicated they already use flash technology, and another 38% have made plans to or are currently investigating the technology.
Last week we published part one of this two-part blog series, “How to Successfully Navigate Enterprise Storage Sizing and Pricing Issues.” This week, we will focus on challenges related to how to think about enterprise storage and how to evaluate it.
Today’s IT leaders are at a crossroads. Behind them, there’s a long legacy of hardware and software deployment decisions that have served the business well for years. But new business requirements and application development methods have begun to test the status quo, and as they seek to modernize, they will face several difficult decisions. IT leaders can either commit to the cloud, go all in with on-premises infrastructure or evaluate something entirely new.
Despite the growing popularity of cloud-based workloads, many companies’ own enterprise data centers keep thriving and growing. Among these on-going investments is the need to upgrade or replace an organization’s current enterprise storage. Often considered the central cog in a data center network, enterprise storage plays a key custodial role in housing many organizations’ mission-critical data assets.
One of the biggest struggles with managing an enterprise data center is the need for various tools with multiple interfaces to manage the different systems associated with IT. This struggle is compounded with the fact that each of these data center systems do not talk to each out of the box, and complex integrations begin to take over. HPE Synergy addresses this challenge by delivering an infrastructure that can manage the technical, as well as the organizational side by combining storage, compute, and network equipment into one.
Despite the growing popularity of cloud-based storage, many enterprises have seen that their data centers keep thriving and growing. Companies that seek increased governance, security and protection of their data continue to invest in their own on-prem data center environments, especially when it comes to storage.
Many organizations are intrigued by the concept of Disaster Recovery as a Service (DRaaS). The biggest lure? You may no longer have to pay capital costs to set up and staff a secondary data center in order to recover systems after a disaster. In the days before cloud, having dual data center sites was one of the few ways to ensure rapid recovery of systems after a disaster. However, due to its cost, it was an option typically reserved for large companies or those in highly regulated fields. Disaster Recovery as a Service now makes secondary storage available to many small-to-midrange organizations, and what’s more, DRaaS providers offer many different variations on the theme of cloud-based recovery. [click to tweet]
Hybrid IT and flash storage are an impressive match. With hybrid IT’s structure combining a mixture of on and off premise resources, it delivers services in a more optimized, balanced and automated fashion, while flash storage has the ability to offer agility and impressive backup and disaster recovery capabilities. The two together can provide:
What storage solution can be found in both the enterprise and personal cell phones? Flash storage. It’s a flexible and compact option that does away with the traditional Hard Disk Drive (HDD). While it is becoming more common to see flash in both the enterprise as well as consumer electronics, the similarities stop there. In this article, you will discover how HPE’s Nimble flash storage is not just any other solution on the market and why you should choose their secondary flash array.
Think about how much data your company created in 2016. According to IDC, the world collectively created about 16.3 zettabytes (yes, a zettabyte is a billion terabytes). They also predict that the world will create 10x that amount by the year 2025. Approximately 90% of that data will be stored in file and object storage. While consumers have traditionally created the bulk of the data up to now, enterprises will create 60% of the world’s data in 2025. At that time, ten percent of all data will be created by IoT. This brings with it a number of challenges as a result of this exponential growth.
Data plays a critical role driving decisions today for your company and your competitors. Having the ability to access your data as quickly and efficiently as possible can provide a competitive edge in a crowded and disruptive marketplace. The demand to acquire the data you need, when you need it, is why many companies are turning to all-flash storage systems that are smart, fast and efficient. Your enterprise storage system is about more than just storing your company data. It is about making your data work for your business and the customers it serves.
A hyperconverged infrastructure provides enterprises with the ability to expand their network into the virtual realm. This setup combines the functionality of traditional hardware into manageable software functions, all of which are controlled through one tool. This ability to use services as a software allows enterprises to respond more quickly to internal and external environmental changes, plus it offers the security, capacity, and customization ability needed to be successful.
Those who are pursuing the latest architype known as hybrid IT in order to revolutionize their enterprise and complete the digital transformation of their organizations know the dream. The dream is to free ourselves from the isolated silos of the traditional data center and obtain the flexibility of a free flowing ecosphere in which workloads are matched with the right platform that both optimizes the user experience and maximizes ROI at the same time. Hybrid IT is about obtaining an unparalleled degree of elasticity in order to migrate applications at will amongst hosting structures (cloud and on-premises). This agility not only delivers infrastructure plasticity, but peace of mind as well by delivering on the ability to achieve near absolute levels of business continuity and disaster recovery.
It is a truly amazing world in which we work today. While the flow of capital is still essential for businesses to grow, materialize, and sustain them, it is ideas today that feed the global economy. Today’s businesses rely on innovators and visionaries. The delivery method for these ideas and revolutionary concepts is IT. IT is a part of nearly every business today because they depend on one or more essential apps to either communicate with their customers or manage their business. Just as the flow of capital is attracted to the most fluid and efficient markets, customers, and business transactions, tech stacks are attracted to those businesses with the most innovative, efficient, and flexible apps and technology. Those that can simplify and streamline the transaction experience will have an advantage over their competitors. In order for your business to make the most of any web related advantage, you need an adaptable ecosphere that is conducive to the constant evolution of those apps through DevOps.
Imagine for a moment that you are calling technical support for a traditional storage solution. You are first greeted by a customer service representative whose job is to take basic information about your problem at hand and forward it on to the appropriate technical support technician or engineer. The representative will ask for the usual product ID numbers, your name, contact information, and remind you of the expiration date of your current service contract. Once your customer profile is established, the barrage of questions begins:
For automobile owners, nothing probably generates as much uncertainty as the “Check Engine” light on the dashboard. A stream of questions commences through your mind once that ominous amber light announces its presence.
If your data center is on an evolutionary track from siloed and hardware-centric to agile and software-defined, you’re aware of converged and hyperconverged infrastructures. If you haven’t yet been introduced to composable infrastructure, welcome to the next gen step in your data center modernization journey.
You’ve probably heard talk about virtualized desktop infrastructure (VDI) and Desktop as a Service (DaaS), but do you know what benefits it can offer your organization? While desktop virtualization isn’t a new concept, amazingly, these technologies have the power to improve user productivity and allow your team members to accomplish more at your enterprise. Read on for a look at VDI and DaaS, as well as ways you can put it to use at your company.
If your data center is on an evolutionary track from siloed and hardware-centric to agile and software-defined, you’re aware of converged and hyperconverged infrastructures. If you haven’t yet been introduced to composable, welcome to the next gen step in your data center modernization journey.
For IT managers of today, the technology world is a constantly changing place, with many new changes popping up on an almost daily basis. Some of these disruptive new technologies include the expansion of advanced virtualization, new cybersecurity monitoring tools and the emergence of new cloud service delivery models, from a wide range of providers. Because of these disruptions in the IT world, the need for a robust backup and recovery strategy is greater than ever before.
For IT managers, the days of “just” keeping the data center up and running are about over. In IT today, it’s no longer just about managing and maintaining assets, and providing support for the back office. In this blazing-fast digital age where collaboration is king, IT is in a better position now more than ever to help drive key business initiatives and help businesses meet strategic goals.
Today’s IT Manager has to walk a tightrope. Management is saddled with the inherited role of supporting the traditional data center that remains built around an inflexible hardware-based infrastructure. At the same time, management and market pressures are compelling them to try to and transform this rigid environment into a modern data center—designed around flexibility, operational velocity and borderless adaptation. It is a quandary of duality that IT teams frustratingly have to deal with.
Many organizations are investigating a “Cloud First” approach with their applications; to save on costs of keeping physical hardware, they want to offload the majority of their IT infrastructures to one or more external providers. While that prospect may fully come to fruition at some time in the future, applications and cloud technologies still have much to develop and change before most organizations will be ready for such a wholesale move.
In today’s data-driven world, nearly all businesses are conducting activities in a digital manner and can’t afford to lose their critical data and information. There are many ways a business continuity interruption can strike, including a malware or ransomware attack, human error or accident, a malicious insider or even a natural disaster. Instead of scrambling to pick up the pieces after such an unfortunate event, it’s best to create a solid disaster recovery (DR) strategy now.
In 2016, network security has been a topic of concern entering conversations in enterprise boardrooms across the world. With many recent high-profile hacks, security breaches and incidents of malware and ransomware, it’s no wonder organizations are quickly seeking strategies to beef up their security and avoid threats. Read on for five tips for improving enterprise network security.
Organizations are currently faced with a cloud computing dilemma: should you use big data solutions or stick with the traditional data warehouse? If you choose the wrong platform to handle your company’s workload, you may find yourself shelling out hundreds or thousands of dollars in frivolous fees. Let’s take a look into what big data and the data warehouse can offer to help you determine which option is right for your organization.
Veeam’s new Availability for the Modern Data Center doesn’t just offer the opportunity for your company to adopt an “always on” business model; it offers many solutions and services traditional legacy systems like Symantec NetBackup do not. In this blog post we will cover the top 10 reasons to consider making the switch when comparing cloud services for backup solutions.
It’s important to address common challenges your business may face when transitioning to a model where you’re “always on” as opposed to being available to customers only during office hours. When comparing cloud services, you’ll want to find a model that offers you flexibility and consistency.