Virtualization Featured Article


VMTurbo, Pure Storage Efforts May Improve SLA Efficacy

Share
Tweet
February 19, 2015

VMTurbo, the creator of the Demand-Driven Control platform, which allows businesses to manage their cloud and virtualized networks, recently announced that it has paired up with Pure Storage, a vendor of solid-state arrays, to give their joint enterprise clients better control over their service-level agreements.

The combination of the VMTurbo's platform and Pure Storage's arrays will allow enterprises to track the activity of virtual workloads within the Purity Operating Environment that sits atop all Pure Storage products. Specifically, VMTurbo's platform will discover behavioral and diagnostic information about network workloads in order to find those workloads, which are overburdened or producing latency that affects entire arrays. The end result will assure that service providers can deliver the quality of service that their SLAs state they will deliver to their customers.

Geeta Sachdev, the CMO of VMTurbo, spoke about the combination of efforts of these two companies and the market trend toward the adoption of flash arrays.

“We are ecstatic to offer our first all-flash array integrated rendition of Storage Control,” Sachdev said. “The combination of Pure Storage and VMTurbo in the marketplace offers a valuable and compelling solution which permits joint customers unprecedented control over application SLAs as all-flash adoption accelerates.”

Similarly, Matt Kixmoeller, the vice president of products at Pure Storage, noted that the control Sachdev mentions directly relates to the enforcement and fulfillment of SLAs. Service providers need to be able to live up the the expectations set forth in their service agreements, and granular control over the activity within their flash arrays will help them get there. Furthermore, the usage of this combination of products can help enterprises lower their costs because arrays can handle more density, regarding the number of workloads they manage in one place, and can operate more efficiently by eliminating the need to spread workloads across a greater number of servers.

A Research and Markets report from last year said the demand for flash storage is expected to expand the market from $500 million in 2014 to $1.6 billion by 2016. If that prediction holds true, it would represent a compound annual growth rate of nearly 60 percent. Technological improvements in how flash storage operates are expected to lower costs and make such storage options available to enterprises that want high performance without paying a premium.




Edited by Maurice Nagle

Article comments powered by Disqus


Freedom from rigid architectures
Learn More ›
FREE Transforming Network Infrastructure eNewsletter - Sign Up

Featured Blog Entries

What Fiber Mountain's Interop Recognition Means for Our Industry

When Fiber Mountain™ began its journey with a launch at Interop New York last fall, we certainly believed that we had a solution that would make a significant impact in the data center space.

What On-Board Optics Means for Density and Flexibility

This past week I read an article in Lightwave Magazine and another in Network World about the formation of the Consortium for On-board Optics (COBO), a group that seeks to create specifications and increase the faceplate density of data center switches and adapters.

Scaling Hyperscale in an Age of Exponential Growth and Virtualization

Over the past several years server, network, storage and application virtualization has revolutionized the way hyperscale data centers are built by consolidating workloads. The trend has simplified network architecture significantly and resulted in huge cost savings as well.

SDN can be the "GPS" Data Center Networks Need

Almost 30 years ago, I came to the USA to attend college, and in my early years as a student I spent every winter, spring and summer break traveling to different parts of this beautiful country.

How Fiber Mountain Future-Proofs Your Data Center

By now you most likely noticed that one of the topics I focus on continually is the problem data centers face today in meeting bandwidth needs. Until now, data centers were forced to purchase fire-breathing, million-dollar core switches to handle the growing volume of traffic, a solution that is both expensive and inefficient.

Video Showcase