Red Hat reveal big data plans, open sources HDFS replacement
JBoss parent company plans to build robust network of partners instead of rolling own Hadoop distribution.
Hat will be open-sourcing their Red Hat Storage Hadoop plug-in to
the Apache Software Foundation, they announced yesterday as part of
reveal of their larger plans in the big data space.
Rather than becoming a Hadoop vendor themselves, Red Hat are
planning to leverage their popular Linux distribution and
middleware solutions to jump on board the big data bandwagon.
In addition, the company says it is building a “robust network of
ecosystem and enterprise integration partners”, although who these
partners will be and how they will “allow users to integrate and
install comprehension enterprise big data solutions” is as yet
Greg Kleiman, director of business strategy for Red Hat’s storage
business unit, told
The Register that they considered developing their own Hadoop
distribution, but decided not to “at the moment”.
“We are going to work through the existing Hadoop distributors,”
said Kleiman. “We are going to play the field and see what
Red Hat Storage Server (previously known as Gluster before Red
Hat’s acquisition) can serve as a direct replacement for Hadoop’s
native HDFS, although the underlying technology has some
differences, particularly in its handling of nodes.
The open-source Hadoop plug-in will be submitted to the Apache
Software Foundation, allowing others to potentially integrate Red
Hat’s storage technology into their own distributions of Hadoop. In
addition, the company is working on a connector between Hive,
Hadoop’s data warehouse system, and its JBoss middleware.
In the accompanying press release, an IDC analyst is quoted as
saying that the big data market is likely to grow from “$6 billion
in 2011 to $23.8 billion in 2016” – so it’s no wonder that Red Hat
has finally made a move for a slice of the Hadoop-flavoured