Apache Hadoop is an open source software platform for distributed storage and distributed processing of very large data sets on computer clusters built from commodity hardware. Hadoop services provide for data storage, data processing, data access, data governance, security, and operations.
The genesis of Hadoop came from the Google File System paper that was published in October 2003. This paper spawned another research paper from Google – MapReduce: Simplified Data Processing on Large Clusters. Development started in the Apache Nutch project, but was moved to the new Hadoop subproject in January 2006. The first committer added to the Hadoop project was Owen O’Malley in March 2006. Hadoop 0.1.0 was released in April 2006 and continues to be evolved by the many contributors to the Apache Hadoop project. Hadoop was named after one of the founder’s toy elephant.
In 2011, Rob Bearden partnered with Yahoo! to establish Hortonworks with 24 engineers from the original Hadoop team including founders Alan Gates, Arun Murthy, Devaraj Das, Mahadev Konar, Owen O’Malley, Sanjay Radia, and Suresh Srinivas.
Some of the reasons organizations use Hadoop is its’ ability to store, manage and analyze vast amounts of structured and unstructured data quickly, reliably, flexibly and at low-cost.
The Hadoop Distributed File System (HDFS) provides scalable, fault-tolerant, cost-efficient storage for your big data lake. It was designed to span large clusters of commodity servers scaling up to hundreds of petabytes and thousands of servers. By distributing storage across many servers, the combined storage resource can grow linearly with demand while remaining economical at every amount of storage.
MapReduce is the original framework for writing massively parallel applications that process large amounts of structured and unstructured data stored in HDFS. MapReduce can take advantage of the locality of data, processing it near the place it is stored on each node in the cluster in order to reduce the distance over which it must be transmitted.
More recently, Apache Hadoop YARN opened Hadoop to other data processing engines that can now run alongside existing MapReduce jobs to process data in many different ways at the same time, such as Apache Spark. YARN provides the centralized resource management that enables you to process multiple workloads simultaneously. YARN is the foundation of the new generation of Hadoop and is enabling organizations everywhere to realize a modern data architecture.
Apache Tez is an extensible framework for building high performance batch and interactive data processing applications, coordinated by YARN in Apache Hadoop. Tez improves the MapReduce paradigm by dramatically improving its speed, while maintaining MapReduce’s ability to scale to petabytes of data.
Applications can interact with the data in Hadoop using batch or interactive SQL (Apache Hive) or low-latency access with NoSQL (Apache HBase). Hive allows business users and data analysts to use their preferred business analytics, reporting and visualization tools with Hadoop. Data stored in HDFS in Hadoop can be searched using Apache Solr.
The Hadoop ecosystem extends data access and processing with powerful tools for data governance and integration including centralized security administration (Apache Ranger) and data classification tagging (Apache Atlas), which combined enable dynamic data access policies that proactively prevent data access violations from occurring. Hadoop perimeter security is also available to integrate with existing enterprise security systems and control user access to Hadoop (Apache Knox).
A very common request from many customers is to be able to index text in image files; for example, text in scanned PNG files. In this tutorial we are going to walkthrough how to do this with SOLR. Prerequisites Download the Hortonworks Sandbox Complete the Learning the Ropes of the HDP Sandbox tutorial. Step-by-step guide […]
Overview The Azure cloud infrastructure has become a commonplace for users to deploy virtual machines on the cloud due to its’ flexibility, ease of deployment, and cost benefits. In addition, Microsoft has expanded Azure to include a marketplace with thousands of certified, open source, and community software applications, developer services, and data—pre-configured for Microsoft Azure. […]
In this tutorial we will explore how you can use policies in HDP Advanced Security to protect your enterprise data lake and audit access by users to resources on HDFS, Hive and HBase from a centralized HDP Security Administration Console.
Introduction Apache Ranger delivers a comprehensive approach to security for a Hadoop cluster. It provides a central security policy administration across the core enterprise security requirements of authorization, accounting and data protection. Apache Ranger already extends baseline features for coordinated enforcement across Hadoop workloads from batch, interactive SQL and real–time in Hadoop. In this tutorial, […]
Overview Apache Ambari is a completely open operational framework for provisioning, managing and monitoring Apache Hadoop clusters. Ambari includes an intuitive collection of operator tools and a set of APIs that mask the complexity of Hadoop, simplifying the operation of clusters. In this tutorial, we will walk through the some of the key aspects of […]
Introduction Hortonworks has recently announced the integration of Apache Atlas and Apache Ranger, and introduced the concept of tag or classification based policies. Enterprises can classify data in Apache Atlas and use the classification to build security policies in Apache Ranger. This tutorial walks through an example of tagging data in Atlas and building a […]
This tutorial will help you get started with Hadoop and HDP. We will use an Internet of Things (IoT) use case to build your first HDP application.
Apache Spark is a fast, in-memory data processing engine with elegant and expressive development APIs in Scala, Java, Python, and R that allow data workers to efficiently execute machine learning algorithms that require fast iterative access to datasets (see Spark API Documentation for more info). Spark on Apache Hadoop YARN enables deep integration with Hadoop […]
“Banking as a service has long sat at the heart of our economy. In our digitally enabled world, the need to seamlessly and efficiently connect different economic agents who are buying and selling goods and services, is critical. The Open Banking Standard is a framework for making banking data work better: for customers; for businesses […]
Wow, I really can’t believe it has only been one year since we launched Hortonworks Community Connection — HCC. What started as a project to make communication between our technical teams more transparent has blossomed into a fantastic and engaging website. Here are just some of the interesting numbers: There are now over 40,000 assets […]
The first blog in this two part series (Deter Financial Crime by Creating an effective AML Program) described how Money Laundering (ML) activities employed by nefarious actors (e.g drug cartels, corrupt public figures & terrorist organizations) have gotten more sophisticated over the years. Global and Regional Banks are falling short of their compliance goals despite huge […]
It is that time of year again, right before Christmas in Las Vegas, where nearly 30,000 technologists gather to see the latest in innovation around the Cloud. Hortonworks is honored to participate as an exhibitor for the first time. If you are in Vegas this week for the AWS re:Invent, please stop by our booth #2732 […]
The first post in this three part series on Digital Foundations @ http://www.vamsitalkstech.com/?p=2517 introduced the concept of Customer 360 or Single View of Customer (SVC). We discussed specific benefits from both a business & operational standpoint that are enabled by SVC. This second post in the series introduces the concept of a Customer Journey. The third & final […]
As discussed in the previous blog in this series @ http://hortonworks.com/blog/frtb-fundamental-review-trading-book-changes-banking-risk-management/, the FRTB (Fundamental Review of the Trading Book) compels Banks to create unified teams from various departments – especially Risk, Finance, the Front Office (where trading desks sit) and Technology to address all of the above significant challenges of the regulation. From a technology capabilities standpoint, the FRTB […]
Earlier this year, we started making Technical Previews of Hortonworks Data Cloud for AWS available. The feedback and response has been incredible, and over the past few months, we performed many Technical Preview refreshes. Now we are ready to make it official and release the product into AWS Marketplace. Therefore, we are excited to announce […]
Regulatory Risk Management evolves… The Basel Committee of supranational supervision was put in place to ensure the stability of the financial system. The Basel Accords are the frameworks that essentially govern the risk taking actions of a bank. To that end, minimum regulatory capital standards are introduced that banks must adhere to. The Bank of International Settlements […]
We were really excited to welcome a sold out crowd at the first Hadoop Summit in Tokyo last week. This was a fantastic response, based on the huge interest around a technology that is transforming industries across Asia and Pacific. We could not put this kind of conference on without the help of our sponsors […]
Enterprises are outgrowing their traditional business intelligent solutions and struggling to use the data they have collected to gain real insights. What happens is that data engineers will scramble to shift data sets between repositories so that data can be analyzed using older methods. As data volumes and data sources grow this problem is only getting […]
THE STATE OF GLOBAL FINANCIAL SERVICES IT ARCHITECTURE… This blog has time & again discussed how Global, Domestic and Regional banks need to be innovative with their IT platform to constantly evolve their product offerings & services. This is imperative due to various business realities – the increased competition by way of the FinTechs, web scale players delivering […]
I started my journey at Hortonworks a little over five years ago and have been to many Hadoop Summits in the US, Europe and now Asia. I just kicked off a two day sold out show in Tokyo and shared my thoughts and stories on architectures and business transformation. John Kreisa’s post here has […]
Last week we announced third quarter results, and it was a milestone quarter. Customers in 60 countries chose Hortonworks to help with their on-premise, hybrid and public cloud data management strategies. A big thanks to the Hortonworks team. I wanted to stop for a second and note the significance of the fact that we just […]
This three part series explores the automated investment management or the “Robo-advisor” (RA) movement. The first post in this series @- http://www.vamsitalkstech.com/?p=2329 – discussed how Wealth Management has been an area largely untouched by automation as far as the front office is concerned. As a result, automated investment vehicles have largely begun changing that trend and they helping create a […]
One of the privileges of my job is working with many different technology leaders who are major users of big data. In all types of company and industry. I get a bird’s eye view of the playing field. There’s definitely a number of skirmishes happening at the moment in the cultural revolution created by the data […]
If you’ve been reading along, you’re aware of the lightbulb moments from my article, “echo: hello world”, that allowed me to discover the benefits of an analytic approach to cybersecurity. This is the start of my new blog series, The CISO’s View, where I focus on the executive level business concerns facing security leaders. Today, […]
Apache Spark has been Open Source’s new kid on the block. Companies are using Spark to develop sophisticated models that would enable them to discover new opportunities or avoid risk. But what does the future or at least the near future hold for Spark? In this blog we have outlined five trends we see in […]
“Millennials want “finance at their fingertips”..they want to be able to email and text the financial advisors and talk to them on a real-time basis,”– Greg Fleming, Ex-Morgan Stanley Executive The first post in this series on Robo-advisors, touched on the fact that Wealth Management has been an area largely untouched by automation as far as […]
I am super excited to be in my first month at Hortonworks, heading up the product and solutions marketing team. In addition to joining a super star team, I am joining one of the leading innovators in the modern data landscape. My love affair with all things data started in the early 1990’s when I […]
Hortonworks DataFlow has been seeing great success being deployed in multiple use cases. We recently shared a set of real-world use cases on a webinar, and also wanted to share here so readers can peruse which types of uses cases are being implemented and see if there are parallels between current and future users of […]
Wealth Management is the highest growth businesses for any medium to large financial institution. It also is the highest customer touch segment of banking and is fostered on long term (read extremely lucrative advisory) relationships. This three part series explores the automated “Robo-advisor” movement in the first post. We will cover the business background and some definitions . The second post will focus on the […]
It’s no secret that there is a data explosion. A recent IDC analyst report from April 2014 indicated the volume of data, known as the digital universe, is doubling in size every two years. And by 2020, there will be as many digital bits as there are stars in the universe. There are many reasons […]
Guest author: Jeff Kelly, Data Strategist, Pivotal The phrase “digital transformation” gets bandied about a lot these days, but what exactly does it mean? When you strip away the hyperbole, I believe digital transformation is the process by which enterprises evolve from using traditional information technology to merely support existing business models to adopting modern […]
Provenance, Lineage & Chain of Custody The models of Provenance, Lineage and Chain of Custody are used in fine art to determine when a piece was created, the sequence of locations where it was held, how it was touched along the way, and who has owned it since creation, all with the purpose of authenticating the piece. […]
The first post in this three part series on Digital Foundations introduces the concept of Customer 360 or Single View of Customer (SVC). We will discuss the need for & the definition of the SVC as part of the first step in any Digital Transformation endeavor. We will also discuss specific benefits from both a […]
Apache, Hadoop, Falcon, Atlas, Tez, Sqoop, Flume, Kafka, Pig, Hive, HBase, Accumulo, Storm, Solr, Spark, Ranger, Knox, Ambari, ZooKeeper, Oozie, Metron and the Hadoop elephant and Apache project logos are either registered trademarks or trademarks of the Apache Software Foundation in the United States or other countries.