The Hortonworks Blog

Posts categorized by : Business Value of Hadoop

Talend is a Hortonworks Certified Technology Partner, and our guest blogger today is Shawn James, director, big data business development, Talend. Shawn and Jim Walker, director of product marketing at Hortonworks, are our guest speakers in an upcoming webinar on Feb. 12th.

If you are a data scientist, MapReduce or Hadoop developer, you are in demand given the massive increase in data science-based projects. These projects are being driven by the private sector of course, but also by a public sector that is looking to tackle a new range of use cases using big data.…

Informatica users leveraging HDP are now able to see a complete end-to-end visual data lineage map of everything done through the Informatica platform. In this blog post, Scott Hedrick, director Big Data Partnerships at Informatica, tells us more about end-to-end visual data lineage.

Hadoop adoption continues to accelerate within mainstream enterprise IT and, as always, organizations need the ability to govern their end-to-end data pipelines for compliance and visibility purposes. Working with Hortonworks, Informatica has extended the metadata management capabilities in Informatica Big Data Governance Edition to include data lineage visibility of data movement, transformation and cleansing beyond traditional systems to cover Apache Hadoop.…

Novetta is an HDP Certified Technology Partner and YARN Certified, delivering agile big data analytics solutions with HDP. In this blog, Jennifer Reed, director of product management at Novetta, shares a recent customer use case in the Oil and Gas vertical market.

Energy companies discover and develop petroleum resources to meet the needs of the global economy. Often, this requires working in areas of the world where organized crime, piracy, and terrorist activity are common and, conversely, where governmental protection is limited.…

This guest blog post is from Alyssa Jarrett, product marketing manager at Splice Machine. Splice Machine is a Hortonworks Certified Technology Partner and provides one of the only Hadoop RDBMS to power a new generation of real-time applications and operational analytics. With its recent Certification with HDP, Splice Machine offers a 10x price/performance improvement over traditional relational databases.

Built on top of the HDFS and Apache HBase components in the Hortonworks Data Platform (HDP), Splice Machine is delighted to announce that it has completed the required integration testing with HDP.…

VoltDB is a Certified Hortonworks Technology Partner and developers of an in-memory relational DBMS capable of supporting high volume OLTP and real-time analytics with Hortonworks Data Platform. Our guest blogger today is John Piekos, vice president of engineering at VoltDB.

It’s a common phrase here at VoltDB: Streaming Apps are Really Database Apps When You Use a Database that’s Fast Enough.

What does that mean?

We’re seeing a trend: developers are struggling to create interactive, real-time applications on fast streaming data.…

Have you ever wondered how to share content infrastructure that transparently synchronizes information with your existing systems? Are you looking for ways to build an open standards-based platform for deep analysis and data monetization? If so, you will want to join our webinar on Wednesday, January 21st, at 10 AM PT.

Our Big Data experts will teach you how to:

  • Leverage 100% of your data, including text, images, audio, video, and many more data types to be automatically consumed and enriched using HP Haven and Hortonworks Data Platform (HDP).
  • Cindy Maike, GM, Insurance at Hortonworks

    Financial services and the insurance industry are projected by many research organizations to benefit significantly from the usage of new information and advanced analytics by blending new data sources. As an industry, insurance depends on data; however, we have always had to struggle with it and now, we have an explosion of new data available.

    The key is to determine what new information is important to your business strategy and what new questions it can provide insights for, and weave that into existing data sources.…

    Our SI partner Ingenious Qube worked with a customer who wanted to price auto insurance based on driving behavior insights obtained from sensors on cars using Hortonworks Data Platform. Rajnish Goswami, CEO of Ingenious Qube, describes the customer story below.

    The Situation

    Insurance companies around the world strive to provide lower insurance rates, and auto insurance is no exception to this phenomenon. The automobile insurance companies are devising ways to derive innovative pricing models that will help customers reduce their insurance premiums; however, it requires an understanding of how one drives their vehicle.…

    As we approach the opening bell on Nasdaq and another milestone for open source Apache Hadoop, we at Hortonworks want to thank those who have contributed deeply to this journey. We owe you – our customers – a huge thank you. Your active collaboration with us in the Apache Hadoop community has greatly impacted the trajectory of this platform for data management and has established a path for how thousands of other enterprises can successfully build a new open data architecture that brings all data under management.…

    Many types of industries are finding new opportunities from an abundance of new types of data stored at scale in Hadoop, combined with Hadoop’s ability to process that data at lower costs than traditional platforms. Apache Hadoop and the Hortonworks Data Platform (HDP) can help enterprises turn what used to be data fumes into high-octane fuel that propels their businesses.

    Sign up for the Hadoop industry solutions email series to find out how Hortonworks customers use Hadoop to solve real-world business challenges.…

    The public sector is charged with protecting citizens, responding to constituents, providing services and maintaining infrastructure. In many instances, the demands of these responsibilities increase while government resources simultaneously shrink under budget pressures.

    How can Intelligence, Defense and Civilian agencies do more with less?

    Apache Hadoop is part of the answer. Within the public sector, Hadoop delivers data-driven actions in support of IT efficiency and good government.

    Download the White Paper

    In one example, the United States Internal Revenue Service had to reduce its auditor headcount due to budget cuts.…

    With Apache Hadoop YARN as its architectural center, Apache Hadoop continues to attract new engines to run within the data platform, as organizations want to efficiently store their data in a single repository and interact with it for batch, interactive and real-time streaming use cases. More and more independent software vendors (ISVs) are developing applications to run in Hadoop via YARN. This increases the number of users and processing engines that operate simultaneously across a Hadoop cluster, on the same data, at the same time.…

    Introduction

    In this 2nd part of the blog post and its accompanying IPython Notebook in our series on Data Science and Apache Hadoop, we continue to demonstrate how to build a predictive model with Apache Hadoop, using existing modeling tools. And this time we’ll use Apache Spark and ML-Lib.

    Apache Spark is a relatively new entrant to the Hadoop ecosystem. Now running natively on Apache Hadoop YARN, the architectural center of Hadoop, Apache Spark is an in-memory data processing API and execution engine that is effective for machine learning and data science use cases.…

    As more organizations consider the cloud as a component of their Apache Hadoop deployments, we can look to our partners for a range of solutions designed to meet these needs. This is the first post in a series on partner solutions available for deploying Hadoop in the cloud. We will build on the Hybrid deployment post with general use cases for Hadoop in a Hybrid cloud. Through our partners we have broad set of options for the cloud available today spanning on-premises, virtual and cloud-based deployments.…

    The successful Hadoop journey typically starts with new analytic applications, which lead to a Data Lake. As more and more applications are created that derive value from the new types of data, an architectural shift happens in the data center: companies gain deeper insight across a large, broad, diverse set of data at efficient scale. They create a Data Lake.

    Cisco and Hortonworks have partnered to build a highly efficient, highly scalable way to manage all your enterprise data in a data lake.…