Get Started with Hadoop on Hortonworks Data Platform 1.1 for Windows

We are excited to release the Hortonworks Data Platform 1.1 for Windows as a Generally Available product. In this blog post, I’m going to outline how to get started with HDP 1.1 for Windows.

HDP for WindowsWith HDP for Windows, you can deploy Apache Hadoop and the HDP stack of components natively on a Windows Server cluster. The HDP for Windows download includes an MSI and remote installation scripts. With these artifacts, you can setup a multi-node Hadoop cluster in either a Workgroup or Active Directory Domain networking configuration. This enables HDP for Windows to be deployed for production use in Windows Data centers.

The best way to get started and evaluate HDP is to set up a single node cluster. We’ve written a quick start guide that walks you through all the pre-requisites and install steps needed to get going. With a single node cluster, you can experience the full functionality of the product – load data into HDFS, execute Hive, Pig and MapReduce jobs, schedule processing workflows through Oozie.

HDP enables seamless integration with the Microsoft BI tool ecosystem. You can explore data in HDFS through the Data Explorer  in Excel. You can query and analyze Hive data in Excel by using the ODBC driver to connect to Hive Server 2. You can import/export data from and to SQL Server through Apache Sqoop.

These integrations enable HDP to become an integral part of your Enterprise Data Architecture, and allow you to utilize the same tools that you are familiar with to interact with HDP.

Learn More. Please take a look at the Hortonworks Documentation to learn more about installing and using HDP 1.1 for Windows.

Tell Us About It. Please visit the HDP 1.1 for Windows Forum to ask questions, get help, provide feedback and hear what others are doing with HDP.

Categorized by :
Hadoop Windows

Leave a Reply

Your email address will not be published. Required fields are marked *

Hortonworks Data Platform
The Hortonworks Data Platform is a 100% open source distribution of Apache Hadoop that is truly enterprise grade having been built, tested and hardened with enterprise rigor.
Get started with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.
Modern Data Architecture
Tackle the challenges of big data. Hadoop integrates with existing EDW, RDBMS and MPP systems to deliver lower cost, higher capacity infrastructure.