SHARE
Facebook X Pinterest WhatsApp

Capgemini Aligns with Pivotal to Construct Big Data Lakes

Three Big Problems Big Data Will Create in 2014 While most of the conversation surrounding Big Data these days is squarely focused on Hadoop, Capgemini is betting that Big Data will require a small pantheon of technologies to be successfully deployed and managed. As a result, Capgemini has inked an extensive alliance with Pivotal, a […]

Written By
MV
Mike Vizard
Dec 23, 2013
Slide Show

Three Big Problems Big Data Will Create in 2014

While most of the conversation surrounding Big Data these days is squarely focused on Hadoop, Capgemini is betting that Big Data will require a small pantheon of technologies to be successfully deployed and managed.

As a result, Capgemini has inked an extensive alliance with Pivotal, a unit of EMC, under which it will co-develop a suite of Big Data applications.

According to Steve Jones, Capgemini’s director of strategy for Big Data and analytics, Capgemini has a long history of working with EMC that is now being extended to include the suite of Big Data products offered by Pivotal, which in addition to a distribution of Hadoop includes the Greenplum massively parallel database and the GemFire in-memory database.

Jones says Capgemini is working with Pivotal to develop these applications because the lakes of Big Data that IT organizations will eventually build will be based on multiple types of databases that will need to be tightly integrated with data management and governance tools. In effect, these “data lakes” will then become the central repository through which all other applications access data. Rather than standardizing on Hadoop, however, Capgemini envisions a world where Hadoop provides a common data substrate that is accessed by a variety of database engines running different classes of enterprise applications.

In that construct, Jones says Capgemini expects SQL to remain the dominant language through which business applications access all that data. But the backend data warehouse where all the structured and unstructured data is maintained and governed is going to be much more of a virtual entity.

As IT organizations move deeper into Big Data in the coming year, it’s pretty clear that it will be a lot more involved than setting up a Hadoop cluster. While the degree to which any organization will choose to wade into a proverbial Big Data lake will vary, the thing to keep in mind is not so much how wide the lake is, but rather how deep.

MV

Michael Vizard is a seasoned IT journalist, with nearly 30 years of experience writing and editing about enterprise IT issues. He is a contributor to publications including Programmableweb, IT Business Edge, CIOinsight and UBM Tech. He formerly was editorial director for Ziff-Davis Enterprise, where he launched the company’s custom content division, and has also served as editor in chief for CRN and InfoWorld. He also has held editorial positions at PC Week, Computerworld and Digital Review.

Recommended for you...

Observability: Why It’s a Red Hot Tech Term
Tom Taulli
Jul 19, 2022
Top GRC Platforms & Tools in 2022
Jira vs. ServiceNow: Features, Pricing, and Comparison
Surajdeep Singh
Jun 17, 2022
IT Business Edge Logo

The go-to resource for IT professionals from all corners of the tech world looking for cutting edge technology solutions that solve their unique business challenges. We aim to help these professionals grow their knowledge base and authority in their field with the top news and trends in the technology space.

Property of TechnologyAdvice. © 2025 TechnologyAdvice. All Rights Reserved

Advertiser Disclosure: Some of the products that appear on this site are from companies from which TechnologyAdvice receives compensation. This compensation may impact how and where products appear on this site including, for example, the order in which they appear. TechnologyAdvice does not include all companies or all types of products available in the marketplace.