The Hortonworks Blog

Posts categorized by : Hadoop

Apache HBase is the online database natively integrated with Hadoop, making HBase the obvious choice for applications that rely on Hadoop’s scale and flexible data processing. With the Hortonworks Data Platform 2.2, HBase High Availability has taken a major step forward, allowing apps on HBase to deliver 99.99% uptime guarantees. This blog takes a look at how HBase High Availability has improved over the past 12 months and how it will improve even more in the future.…

We are living in a hyper connected world. Digitization has lead to massive improvements in human productivity and enabled us to find solutions that would otherwise be simply impossible. Spurring digitization has been a perfect confluence of network, compute and analytics. Thanks to cloud computing, individuals and enterprises of any scale can continuously collect & process data using dynamic compute resources. Advanced scale out analytics has enabled enterprises to derive insight and operationalize them for improved outcomes.…

The Beginning of our Oil and Gas Journey

Hortonworks began working with the Oil & Gas industry in November of 2013 and our involvement accelerated during a very busy 2014 campaign. Our momentum was set against a backdrop early in the year of milestones in drilling and production across unconventional shale plays in North America, along with with a number of acquisitions, mergers, and divestitures that continued to shape the industry landscape.…

Have you ever wondered how to share content infrastructure that transparently synchronizes information with your existing systems? Are you looking for ways to build an open standards-based platform for deep analysis and data monetization? If so, you will want to join our webinar on Wednesday, January 21st, at 10 AM PT.

Our Big Data experts will teach you how to:

  • Leverage 100% of your data, including text, images, audio, video, and many more data types to be automatically consumed and enriched using HP Haven and Hortonworks Data Platform (HDP).
  • The Hadoop Distributed File System (HDFS) is the reliable and scalable data storage core of the Hortonworks Data Platform (HDP). In HDP, HDFS and YARN combine to form the distributed operating system for your data platform, providing resource management for diverse workloads and scalable data storage for the next generation of analytical applications.

    In this blog, we’ll describe the key concepts introduced by Heterogeneous Storage in HDFS and how they are utilized to enable key tiered storage scenarios.…

    Last year on December 11th, Hortonworks presented the sixth of 8 Discover HDP 2.2 webinars: Apache HBase with YARN & Slider for Fast NoSQL Access. Justin Sears, Carter Shanklin and Enis Soztutar hosted this 6th webinar in the series.

    After Justin Sears set the stage for the webinar by explaining the drivers behind Modern Data Architecture (MDA), Carter Shanklin and Enis Soztutar introduced Apache HBase and discussed how to use it with Apace Hadoop YARN and Apache Slider for fast NoSQL access to your data.…

    Our SI partner Ingenious Qube worked with a customer who wanted to price auto insurance based on driving behavior insights obtained from sensors on cars using Hortonworks Data Platform. Rajnish Goswami, CEO of Ingenious Qube, describes the customer story below.

    The Situation

    Insurance companies around the world strive to provide lower insurance rates, and auto insurance is no exception to this phenomenon. The automobile insurance companies are devising ways to derive innovative pricing models that will help customers reduce their insurance premiums; however, it requires an understanding of how one drives their vehicle.…

    This guest blog post is from Srikanth Venkat, director of product management at Dataguise, a Hortonworks security partner.

    Plus ça change, plus c’est la même chose As Jean-Baptiste Alphonse Karr noted “The more things change, the more they stay the same.” Often, that’s not what we hear when looking at Hadoop security: people tend to call out how different Hadoop is, and how different its security solutions need to be.…

    With YARN as its architectural center, Apache Hadoop continues to attract new engines to run within the data platform, as organizations want to efficiently store their data in a single repository and interact with it for batch, interactive and real-time streaming use cases. As more data flows into and through a Hadoop cluster to feed these engines, Apache Falcon is a crucial framework for simplifying data management and pipeline processing.

    Falcon enables data architects to automate the movement and processing of datasets for ingest, pipeline, disaster recovery and data retention use cases.…

    We take pride in producing valuable technical blogs and sharing it with a wider audience. Of all the blogs published in 2014 on our website, the following were most popular:

  • Improving Spark for Data Pipelines with Native YARN Integration.

    Gopal Vijayaraghavan and Oleg Zhurakousky demonstrate improved Apache Spark, which with the help of the pluggable execution context.

  • HDP 2.2 A Major Step Forward for Enterprise Hadoop

    Tim Hall outlines six months of innovation and new features across Apache Hadoop and its related projects.

  • Introduction

    Apache Ranger provides centralized security for the Enterprise Hadoop ecosystem, including fine-grained access control and centralized audit mechanism, all essential for Enterprise Hadoop. This blog covers various details of Apache Ranger’s audit framework options available with Apache Ranger Release 0.4.0 in HDP 2.2 and how they can be configured.

    The audit framework can be configured to send access audit logs generated by Apache Ranger plug-ins to one or more of the following destinations:

    • RDBMS: MySQL or Oracle
    • HDFS
    • Log4j appender
    Default Value xasecure.audit.is.enabled Setting to enable/disable audit logging in the Ranger plug-in.…

    On December 4th, Hortonworks presented the fifth of 8 Discover HDP 2.2 webinars: Apache Kafka and Apache Storm for Stream Data Processing. Taylor Goetz, Rajiv Onat, and Justin Sears hosted this 5th webinar in the series.

    After Justin Sears set the stage for the webinar by explaining the drivers behind Modern Data Architecture (MDA), Rajiv Onat and Taylor Goetz introduced and discussed how to use Apache Kafka and Apache Storm for stream data processing.…

    With Apache Hadoop YARN as its architectural center, Apache Hadoop continues to attract new engines to run within the data platform, as organizations want to efficiently store their data in a single repository and interact with it for batch, interactive and real-time streaming use cases. Apache Storm brings real-time data processing capabilities to help capture new business opportunities by powering low-latency dashboards, security alerts, and operational enhancements integrated with other applications running in the Hadoop cluster.…

    Hortonworks introduces HDP Operations Ready, HDP Security Ready and HDP Governance Ready certifications to showcase solutions that deeply integrate with enterprise Hadoop.

    Customer adoption of Apache Hadoop continues to accelerate the pace at which the community works to meet the requirements of Enterprise Hadoop. Once the place of HDFS and MapReduce only, the introduction of Apache Hadoop YARN a little over a year ago has unleashed many new ways to get value from a Hadoop cluster.…

    Hortonworks architects vertically integrate the projects within our Hadoop distribution with YARN and HDFS in order to enable HDP to span workloads from batch, interactive, and real time—across both open source and other data access technologies. In HDP 2.2, we deliver work to vertically integrate Apache Storm, Apache Accumulo and Apache HBase so that all of those long-running services run in Hadoop on YARN via Apache Slider.

    The Apache Slider community recently released Apache Slider 0.60.0.…