Startup Focuses on Hadoop-Powered Integration

Loraine Lawson
Slide Show

Looking Ahead at the 2015 Business Intelligence Landscape

Xplenty is officially in the integration business — and it’s using Hadoop’s processing power as its foundation.

The Tel Aviv start-up originally entered the crowded Hadoop-as-a-Service market. That made companies like Amazon and AltiScale its main competitors. But those companies are essentially providing Hadoop tools and the infrastructure, which means you need to know MapReduce coding to use it.

Xplenty differentiated in that it provided a tool for the ETL data processing, including pre-built connectors, and its graphical user interface abstracts the integration, allowing you to simply drag and drop data sources and transformations. That eliminated the need to know how to code with MapReduce — and Xplenty’s customers appreciated that more than anything, it seemed.

After watching how its customer base used the solution, the company decided to shift course and focus on selling its ability to integrate the data and prepare it for analytics. In other words, it’s bypassing the administrators and appealing straight to the data and analytics users.


“It's essentially the same product. We started out positioning ourselves as simply Hadoop-as-a-service because we thought people were interested in using Hadoop as a technology — and add to that the fact that Hadoop is complicated and there are not enough Hadoop skills out there. We wanted to breach that gap between the technology and the user,” said Xplenty co-founder and CEO Yaniv Mor. “What we found out is actually people — especially the data people, the BI people — don't actually care much about technology, they care about the application. They care about what they can do with the technology.”

Users still benefit from the Hadoop, which provides a scalable foundation for the processing engine, of course. But the selling point is the integration — and that’s resonating well with Xplenty’s target audience, he added.

Right now, Xplenty’s 20-company user base is from the online verticals, primarily in gaming and e-commerce. The typical use case is moving data between cloud solutions, often from Amazon to Rackspace or an analytics tool, but the solution can also handle on-premise to cloud as well. The solution can handle data from multiple data stores, including NoSQL, HDFS, SQL, log management software and cloud storage. While the data processing is currently batch-mode MapReduce, the company plans to add Spark and Tez to support more interactive processing some time this year.

Xplenty

Of course, this puts Xplenty in competition with a whole new batch of vendors, such as Pentaho, Talend and integration powerhouse Informatica. All offer integration on Hadoop, but the main difference is that Xplenty is native to and optimized for Hadoop. A 451 report provided by the company notes that DataWorks and Trifacta’s Data Transformation Platform may emerge as potential competitors.

Xplenty offers a full-feature seven-day trial that is restricted only by data size. The company offers three fixed-rate monthly plans based on the size of infrastructure, and customers also pay for the computing power used by time (i.e., two hours would cost less than 24 hours).

Generally, the price ranges from a few hundred to a few thousand dollars, he said. With that, customers also receive consulting help from the company for the initial implementation.

“Data and BI people are spending 80 percent of their time just crunching the data and only 20 percent doing analysis,” Mor said. “We are trying to change that.”

Loraine Lawson is a veteran technology reporter and blogger. She currently writes the Integration blog for IT Business Edge, which covers all aspects of integration technology, including data governance and best practices. She has also covered IT/Business Alignment and IT Security for IT Business Edge. Before becoming a freelance writer, Lawson worked at TechRepublic as a site editor and writer, covering mobile, IT management, IT security and other technology trends. Previously, she was a webmaster at the Kentucky Transportation Cabinet and a newspaper journalist. Follow Lawson at Google+ and on Twitter.



Add Comment      Leave a comment on this blog post
Feb 11, 2015 5:28 AM Ciaran Dynes Ciaran Dynes  says:
Hi I'd like to highlight that Talend has been optimized for Hadoop for a number of years. But it's an on-going investment, where we learn new techniques and approaches all the time. Talend relies 100% on the Hadoop API, for transformation, mapping, machine learning, security, deployment, job scheduling, everything... Talend Big Data Studio generates native MapReduce, Pig and Hive code, which is optimized for Hadoop. We've spent a lot of time running the TPCH benchmark and performance tuning the way we do things for Hadoop. For example we leverage sorting algorithms that work best for the HDFS file-system, as well as replicated joins, in-memory caching, do lookups in mappers vs. reducers and a number of other Hadoop recommendations. We also do the same for Spark, Spark Streaming and Storm. Cheers Reply
Feb 13, 2015 11:00 AM Loraine Lawson Loraine Lawson  says:
Thanks for pointing that out! I may be calling about Spark soon! Reply
Oct 6, 2016 12:23 AM Steve Hawks Steve Hawks  says:
Informatica certification has a great value in the market. This certification helps to achieve a great position and also it helps to prove our knowledge. Informatica institutes in Chennai is useful to learn the real life problems in the difficult situation. With that we can learn as how to solve the problems in an easy way. Reply

Post a comment

 

 

 

 


(Maximum characters: 1200). You have 1200 characters left.

 

null
null

 

Subscribe to our Newsletters

Sign up now and get the best business technology insights direct to your inbox.