SHARE
Facebook X Pinterest WhatsApp

DAS Rules in Hyperscale, For Now

Five Innovations in the Data Storage Landscape At one time, it appeared that direct attached storage (DAS) was heading the way of the dinosaur. Networked storage was where it was at because you could manage a centralized array of disk and tape drives and distribute data to a wide variety of endpoints. Not only did […]

Written By
thumbnail
Arthur Cole
Arthur Cole
Feb 27, 2015
Slide Show

Five Innovations in the Data Storage Landscape

At one time, it appeared that direct attached storage (DAS) was heading the way of the dinosaur. Networked storage was where it was at because you could manage a centralized array of disk and tape drives and distribute data to a wide variety of endpoints. Not only did this provide enormous capacity when required, it fostered a higher level of data interactivity among users as well.

Times have changed, however, and with modular, hyperscale infrastructure on the rise, it appears that keeping storage close to processing resources is proving to be highly advantageous to scale-out database applications, with advanced network fabric technologies providing the collective capacity and data availability requirements of modern workloads.

This is reflected in the variety of solutions hitting the channel aimed at DAS architectures, starting with higher capacity disk drives. Toshiba recently announced that it had hit the 1 TB/square inch plateau on a conventional disk drive platform, ushering in the possibility of a standard drive that tops 3 TB. The company did not offer many details, other than to say it utilizes a conventional Perpendicular Magnetic Recording (PMR) design rather than the emerging Shingled Magnetic Recording (SMR) approach. The company is expecting to release the drive later this year.

DAS has caught on particularly well in large Hadoop clusters where storage and retrieval of Big Data workloads from a centralized repository would be problematic at best. But as Hadoop evolves beyond simple MapReduce queries and the like to take on more complicated tasks, it is questionable whether DAS will be able to keep up, says Infostor’s Paul Rubens. For one thing, DAS lacks key enterprise features like compliance and regulatory controls, multi-site data protection and collaborative workflow integration. Development around these problems is proceeding on a number of fronts, including various Hadoop storage management solutions, as well as advanced SAN/NAS/appliance solutions, virtual or containerized Hadoop instances and cloud-based clustered storage.

Leveraging DAS in scale-out applications is also emerging as a key strategy among traditional data center vendors. Dell has tapped Nutanix for the storage management components of its XC series appliance, incorporating a distributed file system within each node to aggregate DAS resources across the cluster. In this way, storage pools can be accessed by all hosts and the entire environment is outfitted with a fault-tolerant backup and replication system that can be easily expanded by adding more modules to the cluster.

Data Analytics

This works fine for all of your DAS resources, but what if you wanted to incorporate legacy SAN/NAS architectures as well, and by extension the apps and test/dev capabilities they support? A startup called Springpath is promising to do just that with a distributed file system that supports block, file and object storage, plus advanced Hadoop applications. As well, the system supports traditional virtualization and new container-based approaches like Docker, and it does so without sacrificing speed, performance and advanced enterprise features as you would with a standard multi-layered management stack.

No matter how you arrange your storage in scale-out environments, the key element will be adequate networking. Centralized repositories will require wide data paths to shuttle loads in and out of the cluster, while attached solutions will need a high degree of flexibility to ensure all data is available to the processing centers that need it.

There are cost and performance trade-offs to every solution (as usual), and by and large the correct approach won’t center on which technology is superior but which offers the highest level of support to the application(s) at hand.

It is nice to think that in the software-defined universe, any hardware resource can be repurposed for any need, but when it comes to storage in particular, it just isn’t so.

Arthur Cole writes about infrastructure for IT Business Edge. Cole has been covering the high-tech media and computing industries for more than 20 years, having served as editor of TV Technology, Video Technology News, Internet News and Multimedia Weekly. His contributions have appeared in Communications Today and Enterprise Networking Planet and as web content for numerous high-tech clients like TwinStrata, Carpathia and NetMagic.

Recommended for you...

Top Managed Service Providers (MSPs) 2022
Observability: Why It’s a Red Hot Tech Term
Tom Taulli
Jul 19, 2022
Top GRC Platforms & Tools in 2022
Jira vs. ServiceNow: Features, Pricing, and Comparison
Surajdeep Singh
Jun 17, 2022
IT Business Edge Logo

The go-to resource for IT professionals from all corners of the tech world looking for cutting edge technology solutions that solve their unique business challenges. We aim to help these professionals grow their knowledge base and authority in their field with the top news and trends in the technology space.

Property of TechnologyAdvice. © 2025 TechnologyAdvice. All Rights Reserved

Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.