Cloudera DataFlow (CDF), formerly Hortonworks DataFlow (HDF), is a scalable, real-time streaming analytics platform that ingests, curates, and analyzes data for key insights and immediate actionable intelligence.
Cloudera DataFlow (CDF), formerly Hortonworks DataFlow (HDF), is a scalable, real-time streaming analytics platform that ingests, curates, and analyzes data for key insights and immediate actionable intelligence.
Customer Reviews
User in Real Estate
Advanced user of Cloudera Data FlowHortonworks two main pillars are HDP (Hortonworks Data Platform) and HDP (Hortonworks Data Flow). The former applies to the infrastructure required for building and deploying a data lake, and the latter is about ingestion, in batch or realtime.
Both HDP and HDF rely entirely on opensource projects, this is a distinctive point about Hortonworks.
As an open source project collection, it relies strongly on community activity. You still have the option to contract premium consulting or training services.
Altough it is quickly evolving into Data Science tools availability (eg. Tensorflow incorporate in HDP 3), it can be cumbersome from a developer transitioning from a traditional IDE, into the notebook vs. datalake metaphore.
Because of its open source platform. Make sure that it has the right integration into your current data field.
Typically it is used as an enterprise platform. There are very few companies that use it only departmentally. It solved the business problems of maintaining a pure open source Hadoop environment. It also solves for Disaster Recovery and Security. Hadoop was not designed for Security, but with Hortonworks Ranger and Kerberos, you can implement a world class security framework.