Ready to Get Started?Download HDF
Hortonworks DataFlow (HDF) provides the only end-to-end platform that collects, curates, analyzes and acts on data in real-time, on-premises or in the cloud, with a drag-and-drop visual interface. HDF is an integrated solution with Apache Nifi/MiNifi, Apache Kafka, Apache Storm and Druid.
The HDF streaming data analytics platform includes data Flow Management, Stream Processing, and Enterprise Services.
Collect and manipulate data flows securely and efficiently while giving real-time operational visibility, control, and management.
Build streaming analytics applications in minutes to capture perishable insights in real-time without writing a single line of code.Learn More
Manage the HDF and HDP ecosystem with comprehensive management panel for provisioning, monitoring, and governance.Learn More
HDF has full featured data collection capabilities that are streaming data agnostic and integrated with over 220 processors. Data can be collected from dynamic and distributed sources of differing formats, schemas, protocols, speeds and sizes and from types such as machines, geo location devices, click streams, files, social feeds, log files and videos.
With HDF, data collection is no longer a tedious process. You can manage data in full flight with a visual control panel to adjust sources, join and split streams, and prioritize data flow. HDF also can add contextual data to your streams for more complete analysis and insight. The always-on data provenance and audit trails provides security and governance compliance and troubleshooting as necessary in real-time. Integrated with Apache NiFi, MiNiFi, Kafka and Storm, HDF is ready for high volume event processing for immediate analysis and action. Kafka allows differing rates of data creation and delivery while Storm provides real-time streaming analytics and immediate insights at a massive scale.
HDF secures end-to-end data flow and routing from source to destination with discrete user authorization and detailed, real-time visual chain of custody. Use the visual user interface of HDF to encrypt streaming data, route it to Kafka, configure buffers and manage congestion so that data can be dynamically prioritized and securely sent. HDF enables role-based data access that allows enterprises to dynamically and securely share select pieces of pertinent data. HDF can easily deploy flow management and streaming applications in a Kerberized environment without much operational overhead.
HDF includes a complete streaming analytics module, Streaming Analytics Manager (SAM), to build streaming analytics applications that do event correlation, context enrichment, complex pattern matching, analytical aggregations and create alerts/notifications when insights are discovered. SAM makes building streaming analytics easy for application deverlopers, DevOps and business analysts to build, develop, collaborate, analyze, deploy, manage applications in minutes without writing a single line of code. Analysts use pre-built charts to quickly build analysis and create dashboards, while DevOps can manage and monitor the applications performance right out of the box.
HDF includes, Schema Registry, a central schema repository that allows analytics applications to flexibly interact with each other. This enables users to save, edit, or retrieve schemas for the data they need. This also allows easy attachment of schemas to each data without incurring additional overhead for greater operational efficiency. With schema version management, data consumers and data producers can evolve at different rates. And, through schema validation, data quality is greatly improved. A central schema registry also provides for greater governance of how data is used. Schema Registry is integrated with Apache Nifi and HDF Streaming Analytics Manager.
Build analytics applications easily with drag and drop visual paradigm with drop down analytics functions
Analyze quickly with rich visual dashboard and an analytics engine powered by Druid
Operate efficiently with prebuilt monitoring dashboards of system metrics
Eliminate the need to code and attach schema to every piece of data and reduce operational overhead
Allow data consumers and producers to evolve at different rates with schema version management
Store schema for any type of entity or data store, not just Kafka
Get HDF release notes; guides for users, developers and getting started.
The industry’s best support for Apache NiFi, Kafka and Storm in the enterprise. Connect to our team experts to help guide your journey.
Real-world training from the Big Data experts. Available in person or on-demand whenever you need us.