

Having multiple data streams can be hard to process and even harder to manage. It can be almost as cumbersome as sorting through thousands of reams of paper. Luckily, Hortonworks, Inc. has a way to keep enterprise data structured.
Scott Gnau, CTO of Hortonworks, Inc., and Tendu Yogurtcu, GM of Big Data at Syncsort, Inc., talked with George Gilbert (@ggilbert41), host of theCUBE, from the SiliconANGLE Media team, during the Hadoop Summit 2016 about Hortonworks DataFlow (HDF).
HDF helps enterprises manage the flow and distribution of data. As data platforms become more connected, there’s a demand for “all data, all the time,” said Gnau. In order to make that happen, Hortonworks had to make it easy for customers to access their data.
Hadoop, in tandem with Hortonworks, allows users to choose a data source and operate on a single data pipeline. Hadoop’s ease of optimization made it the perfect complementary platform.
HDF acts as a traffic cop directing and managing the many pieces and containers of programming. But in addition to directing the flow of data, it can also expose metadata in Hadoop, which makes it available for joint partners. This means that not only are processes streamlined, but they can also be recreated and improved by others.
It is necessary to have a supportive ecosystem in order to fully utilize HDF. It takes the confusion of multiple datasets and pipelines and makes the information accessible to all developers and analysts in an enterprise.
Hortonworks wants to be on the leading edge of dealing with massive data streams in an open-source environment. It is the “new way of doing things,” said Gnau.
Watch the complete video interview below, and be sure to check out more of SiliconANGLE and theCUBE’s coverage of the Hadoop Summit US.
THANK YOU