The Hortonworks Blog

Back in September, we presented a 3-part webinar series on our collaborations with Red Hat. Close to a thousand registrants and attendees participated and provided rich interaction to our series. The content included an overview of our strategic partnership, demonstrated a couple of demos, and provided tutorials to get you started on your Big Data journey with Red Hat and Apache Hadoop.

In this blog, Kenneth Peeples, JBoss technology evangelist and principal marketing manager for Data Virtualization and Fuse Service Works at Red Hat, recaps the webinar series and offers insights into JBoss Data Virtualization and HDP.…

On October 15 we announced that we would support Apache Hadoop as an Infrastructure as a Service (IaaS) on Microsoft Azure. This made us the first Hadoop vendor to give customers and prospects access to that flexible and scalable cloud infrastructure for their big data deployments.

This guide walks you through using the Azure Gallery to quickly deploy Hortonworks Data Platform (HDP) clusters on Microsoft Azure IaaS.

What you need is:

  • A Microsoft Azure account
  • That’s it!

Exponential increases in data volumes have forced the data architect and analyst to build much larger and distributed data environments — potentially comprised of hundreds, and sometimes even thousands of servers and switches. Scaling to these cluster sizes does not come without challenges in terms of costs, security and integration with existing infrastructure.

The combination of the Hortonworks Data Platform (HDP) and Cisco UCS allows IT departments and business decisions makers to adopt a new, cost-effective management, massively scalable and secure approach to data within the enterprise.…

Hortonworks Data Platform Version 2.2 represents yet another major step forward for Hadoop as the foundation of a Modern Data Architecture. This release incorporates the last six months of innovation and includes more than a hundred new features and closes thousands of issues across Apache Hadoop and its related projects.

Our approach at Hortonworks is to enable a Modern Data Architecture with YARN as the architectural center, supported by key capabilities required of an enterprise data platform — spanning Governance, Security and Operations.…

More and more enterprises are looking to the cloud as a place to handle a variety of their data processing and backup needs. Apache Hadoop lends itself to running in cloud environments because of the alignment around scalability and flexibility for compute and storage offered with today’s cloud infrastructures. Today, we are excited to announce that the Hortonworks Data Platform (HDP) is the first platform to be certified to run on Azure Infrastructure as a Service.…

Apache Hadoop has taken a mission critical role in the Modern Data Architecture (MDA) with the advent of Apache Hadoop YARN. YARN has enabled enterprises to store and process data across many execution engines at a scale that has not been possible earlier. This in turn has made security a crucial component of enterprise Hadoop. At Hortonworks we have broken the problem of enterprise security into four key areas of focus: authentication, authorization, auditing and data protection.…

Since its first deployment at Yahoo in 2006, HDFS has established itself as the defacto scalable, reliable and robust file system for Big Data. It has addressed several fundamental problems of distributed storage at unparalleled scales and with enterprise grade robustness.

As more and more enterprises adopt Apache Hadoop, it is becoming a unified central storage aka Data Lake for all kinds of enterprise data. Many of these storage use cases are for file storage for classic big data applications, where HDFS is the perfect fit.…

Today’s guest blog comes from Matt Davies at Splunk, where he is the Director of Marketing.

You can’t really escape the fact that we’re in the age of the customer. From CRM to the “long tail” to multi-channel to social media brand sentiment to Net Promoter Scores – it is all about customer experience. Big Data has an important part to play – no great revelation there but how do you actually do it?…

If you are heading to New York City for the Strata Conference,October 15-17, 2014, and are interested in learning more about how Apache Hadoop fits into Modern Data Architecture (MDA) along side key enterprise technologies and want to have a chance for some great prizes, then don’t miss our Passport Program.

You will get an opportunity to meet with Big Data ecosystem leaders, see the Hortonworks Data Platform (HDP) in action, and join the conversation with eighteen of our business partners.…

Since our founding over three years ago, a core part of our strategy has been on enabling the enterprise to use Hadoop in the context of their existing technologies via a Modern Data Architecture. From the earliest days of the company when we hired Mitch Ferguson to head our business development efforts, we’ve been working closely with data center ecosystem leaders, large and small, to integrate Hadoop so that it can take it’s place in the next generation data architecture.…

Over the last several months, Oracle and Hortonworks have been working together to bring ETL tools and connectors to the Hortonworks Data Platform (HDP). Hortonworks and Oracle have teamed up to provide comprehensive data integration capabilities and technologies that lay the foundation for a modern data integration architecture, delivering on the promise of big data for customers of all sizes and scale.

We’re happy to announce that the Oracle Data Integrator (ODI) is now certified with the HDP 2.1.…

It’s been just about a year since SAP signed up to be a strategic reseller for the Hortonworks Data Platform and what a year it’s been — we’ve worked on enablement, engineering and marketing efforts.

Enablement and Support Milestones

As part SAP’s Big Data strategy, the SAP sales force can resell the Hortonworks Data Platform (HDP) as a Strategic Reseller. This means that a SAP customer can purchase HDP from SAP using the same purchasing vehicles and agreements they already have in place.…

Last week’s Hortonworks webinar “What’s Possible with a Modern Data Architecture?” featured Greg Girard, program director for omni-channel analytics strategies at IDC Retail Insights and Mark Ledbetter, vice president for industry solutions at Hortonworks. Greg provides targeted, fact-based guidance to retailers for the application of analytics across the enterprise. Mark has more than twenty-five years experience in the software industry with a focus on retail and supply chains.

Many of Greg and Mark’s thoughts from the webinar echo topics also covered in the recent Hortonworks white paper “The Retail Sector Boosts Sales with Hadoop.”

Download White Paper

Greg discussed the most significant drivers of big data initiatives in the retail industry, including customer acquisition, pricing strategies or competitive intelligence.…

At Hortonworks, we are always watching emerging trends in the datacenter to find opportunities for deeper ecosystem integration with Apache Hadoop in simple and intuitive ways. We first partnered with OpenShift by Red Hat earlier this year when we made it possible to call out to Hadoop services from OpenShift via cartridges. You can read more about that solution here. As Enterprise Cloud (e.g. PaaS) offerings have matured to support a broad set of workloads, we’ve had a number of our customers ask about how Hadoop-centered Big Data and PaaS initiatives could work together – particularly in light of Apache Hadoop YARN being the multi-workload resource manager for batch, interactive and real-time workloads on Hadoop.…

Concurrent Inc. is a Hortonworks Technology Partner and recently announced that Cascading 3.0 now supports Apache Tez as an application runtime. Cascading is a powerful development framework for building enterprise data applications on Hadoop and is one of the most widely deployed technologies for data applications, with more than 175,000 user downloads a month. Used by thousands of businesses including eBay, Etsy, The Climate Corp and Twitter, Cascading is the de facto standard in data application development on Hadoop.…