Explosive growth expected for Hadoop
By Mark Dexter
10th May 2012

The market for software related to the Hadoop and MapReduce programming frameworks for large-scale data analysis will jump from $77 million in 2011 to $812.8 million in 2016, a compound annual growth rate of 60.2%, according to a new report released Monday by analyst firm IDC.

Hadoop is an open-source implementation of the MapReduce framework. It is hosted at the Apache Software Foundation along with a number of supporting software projects, including the Hadoop Distributed File System (HDFS) and Pig programming language.

MapReduce and Hadoop are based on the principle of splitting up large amounts of data and then processing the chunks in parallel across large numbers of nodes. It's closely associated with the industry buzzword "big data," which refers to the ever-larger volumes of information, particularly of unstructured form, being generated by websites, social media, sensors and other sources.

Overall, Hadoop has enjoyed a steady stream of interest from commercial analytics and database vendors in recent years, who have begun offering commercial products and services for it.

While "fantastic and largely unsupportable claims have been made" regarding Hadoop and MapReduce's use cases and benefits, "there can be no doubt that it does provide a relatively low-cost means of deriving considerable value from very large collections of unorganized data," IDC analysts Carl Olofson and Dan Vesset wrote in the report.

Therefore, the conditions are right for significant growth in the Hadoop-MapReduce "ecosystem," according to IDC.

This year, "Leading adopters in the mainstream IT world will move from 'proof of concept' to real value," the report states.

However, lack of qualified talent will limit the technology's rise during the next two to three years, it adds.

The coming years will also see a "battle between open source purists, who believe that the core of Hadoop deployment must be based purely on the Apache project code," according to IDC. However, most IT organizations will use a mix of commercial and open-source components in their Hadoop environments, the report adds.

Still, "competition between open source vendors and their closed source counterparts may force lower license fees from the latter group, resulting in somewhat slower software revenue growth than would be the case if open source projects did not represent so large a component of this market space."

Read at Source: Computerworld

Comments

Currently there are no comments. Be the first to post one!

Post Comment

*
*
*

Is Artificial Intelligence wasting our money?

If you are anything like me, you will have become addicted to online shopping. And like me your addiction might have led you to questioning if you are buying things unnecessarily. Asking yourself: Am I wasting my money? So, as... Read More

15 years on: How times have changed…

When I incorporated KDR 15 years ago this month I had a simple mission. I wanted to run the market leading agency in its field placing contract staff with experience of working with the emerging data warehouse tool Kalido. I... Read More

The life of a Business Administration apprentice!

I started my apprenticeship with KDR in the September of 2018, when I was searching for a job I never imagined that I would find a company quite like KDR, from the moment I first stepped through the door everyone... Read More

How can data analytics impact climate change?

Every quarter we produce an e-magazine on issues and topics within the data and analytics world. In our previous issues we have looked at the impact of blockchain , the future of the data warehouse and GDPR . In our next issue... Read More

Where should we send our newsletter?

Close