We have reached an inflection point in the rate of data creation that, unless you are willing to start throwing huge quantities of it away, you simply cannot afford to continue using the same technologies and tools to store and analyze it. The existing data silos – impractical for many reasons beyond pure expense – simply must be consolidated, even if the full picture of exactly how the utility of each piece of data will be maximized is still unknown.
One potential option many businesses have chosen to pursue in the hopes of addressing current business concerns while also maximizing future possibilities and minimizing future risks is building a data lake. With that, however, comes a separate set of challenges and considerations.
Large data volumes drive the need for data lakes. In simple terms, a data lake is a repository for large quantities and varieties of data, both structured and unstructured. The data is placed in a store, where it can be saved for analysis throughout the organization. In this slideshow, Storiant, a cloud storage provider, has identified six tips on how a data lake can reconcile large volumes of data with the need to access it.
The top three data dangers lurking in cloud environments and tips for how to manage data protection and backup in a SaaS-based world. ... More >>
To keep up with machine data growth and avoid costs it traditionally incurs, companies need to combine on-premises storage performance and availability with the elasticity and economics of the cloud. ... More >>
A majority of IT professionals are facing occasional or ongoing storage performance issues as they look to maintain highly virtualized data center environments that can keep up with business growth. ... More >>