The Hortonworks Blog

Posts categorized by : Other

Hadoop can be a great complement to existing data warehouse platforms, such as Teradata, as it naturally helps to address two key storage challenges:

The purpose of this article is to detail some of the key integration points and to show how data can be easily exchanged for enrichment between the two platforms.

As a data integrator who is familiar with RDBMS systems and is new to the Hadoop platform, I was looking for a simple way (i.e.…

Ever since I was a kid, I’ve used memorable movie quotes to help people understand a key point in a way that lightens the mood and generates some laughs. If you’re going to work hard, you gotta have fun, right???

“Don’t make me angry… you wouldn’t like me when I’m angry”

The big data market is rife with aspirational marketing misinformation, which among other things causes customer confusion, slows the path to value, and frankly, makes me a little angry.…

With the growing number of large-scale enterprise deployments of big data, certain limitations have become more apparent bringing to light some weaknesses in this first phase of analytics infrastructures.  Hadoop, clearly a very valuable tool for the collection of unstructured data, poses some challenges that need to be overcome for wide spread successful enterprise adoption.

In our upcoming webinar on Tuesday Feb 19 at 10 am PT, we will address these issues and highlight how to solve them using Hortonworks Data Platform and our partner Actian.…

We cannot wait to see you at the Santa Clara Convention for the next few days! Hortonworks will be one of the sponsors at the conference and will be presenting in various sessions. If you’re going to be around, attend one (or all) of our sessions and remember to stop by Booth #811. We have a nice schedule lined up for you and we hope you can join us!

Attend our sessions

This year’s Strata Santa Clara, Hortonworks will also participate in a number of presentations on all things about data – don’t miss any of them!…

Microsoft and Hortonworks have been working together for over two years now with the goal of bringing the power of Big Data to a billion people. As a result of that work, today we announced the General Availability of HDP 2.0 for Windows with the full power of YARN.

There are already over half a billion Excel users on this planet.

So, we have put together a short tutorial on the Hortonworks Sandbox where we walk through the end-to-end data pipeline using HDP and Microsoft Excel in the shoes of a data analyst at a financial services firm where she:

  • Cleans and aggregates 10 years of raw stock tick data from NYSE
  • Enriches the data model by looking up additional attributes from Wikipedia
  • Creates an interactive visualization on the model

You can find the tutorial here.…

Encryption is applied to electronic information in order to ensure its privacy and confidentiality.  Typically, we think of protecting data as it rests or in motion.  Wire Encryption protects the latter as data moves through Hadoop over RPC, HTTP, Data Transfer Protocol (DTP), and JDBC.

Let’s cover the configuration required to encrypt each of these protocols. To see the step-by-step instructions please see the HDP 2.0 documentation.

RPC Encryption

The most common way for a client to interact with a Hadoop cluster is through RPC.  …

Apache Sqoop is a tool that transfers data between the Hadoop ecosystem and enterprise data stores. Sqoop does this by providing methods to transfer data to HDFS or Hive (using HCatalog). Oracle Database is one of the databases supported by Apache Sqoop. With Oracle Database, the database connection credentials are stored in Oracle Wallet. Oracle Wallet can act as the store of keys and secrets such as authentication credentials. This post describes how Oracle Wallet adds a secure authentication layer for Sqoop jobs.…

Security is a top agenda item and represents critical requirements for Hadoop projects. Over the years, Hadoop has evolved to address key concerns regarding authentication, authorization, accounting, and data protection natively within a cluster and there are many secure Hadoop clusters in production. Hadoop is being used securely and successfully today in sensitive financial services applications, private healthcare initiatives and in a range of other security-sensitive environments. As enterprise adoption of Hadoop grows, so do the security concerns and a roadmap to embrace and incorporate these enterprise security features has emerged.…

This is the first of two posts examining the use of Hive for interaction with HBase tables. The second post is here.

One of the things I’m frequently asked about is how to use HBase from Apache Hive. Not just how to do it, but what works, how well it works, and how to make good use of it. I’ve done a bit of research in this area, so hopefully this will be useful to someone besides myself.…

The Apache Knox community announced the release of the Apache Knox Gateway (Incubator) 0.3.0. We, at Hortonworks, are excited about this announcement.

The Apache Knox Gateway is a REST API Gateway for Hadoop with a focus on enterprise security integration.  It provides a simple and extensible model for securing access to Hadoop core and ecosystem REST APIs.

Apache Knox provides pluggable authentication to LDAP and trusted identity providers as well as service level authorization and more.  …

It’s been a huge couple of weeks for us at Hortonworks HQ. We’ve talked about the GA of Hadoop 2, the subsequent release of Hortonworks Data Platform 2.0, and a little of the future with Apache Storm. We’ve been staggered by the support, goodwill and enthusiasm we’ve seen from you all.

We hope you’re as excited about Hadoop as we are, and we wanted to say thanks to our amazing team, amazing customers, amazing partners and the most amazing community for doing Hadoop with us – THANK YOU.…

Typical delivery of enterprise software involves a very controlled date with a secret roadmap designed to wow prospects, customers, press and analysts…or at least that is the way it usually works.  Open source, however, changes this equation.

As described here, the vision for extending Hadoop beyond its batch-only roots in support of interactive and real-time workloads was set by Arun Murthy back in 2008. The initiation of YARN, the key technology for enabling this vision, started in earnest in 2011, was declared GA by the community in the recent Apache Hadoop 2.2 release, and is now delivered for mainstream enterprises and the broader commercial ecosystem with the release of Hortonworks Data Platform 2.0.…

Today we are proud to announce the delivery of Apache Ambari 1.4.1. Ambari 1.4.1 combines many months of work in the community advancing the Ambari codebase. Over 760 JIRAs have been resolved since the Ambari 1.2.5 release. We would like to thank the nearly 40 engineers who contributed to help make this release possible.

Hello Hadoop 2, Meet Apache Ambari
The most important addition to Ambari 1.4.1 is support for installing, managing and monitoring a cluster based on the Hadoop 2 stack.…

The Hortonworks HBase team is excited to see HBase 96 released.  It represents a broad community effort and massive amount of work that has been building for more than a year.

HBase 96 closes out over 2000 issues (2134 Jira tickets to be exact) and it represented the collective work from a VERY active community. Kudos to everyone involved! As the authors in a recent Apache blog alluded to, the HBase community is very healthy and includes developers from many companies including Hortonworks, Yahoo!, Cloudera, Salesforce, eBay, Intel, and Facebook, just to name just a few.…

Today we announced the Analytics Advantage with Hadoop offering from SAS, Teradata and Hortonworks. The new offering leverages the capabilities for in-database data preparation, analytic model building and deployment and combines Teradata’s Appliance for SAS® High-Performance Analytics offering with the Teradata Appliance for Hadoop built on Hortonworks Data Platform. Using Teradata’s Unified Data Architecture (UDA), this high-speed integrated offering allows customers to discover, build and deploy analytic models across data stored in Teradata and Hadoop, promoting businesses’ ability to act upon analytic insights from any type of data across a seamless environment and faster than ever before.…

Go to page:12345...Last »