Enterprise-Grade Hadoop is a Reality Today
The following is a guest post from Scott Gnau, President, Teradata Labs
I continue to be astonished by the evolution of Apache Hadoop, the software framework for large scale computing that has flourished thanks to a dynamic open source ecosystem. An army of contributors, including the smart engineers and contributors at Hortonworks, constantly refines Hadoop’s ability to manage massive amounts of data on computer clusters via MapReduce processing and the underlying Hadoop Distributed File System (HDFS). Images, documents, web logs, and scientific data join myriad other forms of information for processing data on tasks ranging from search to machine learning, scoring, and data refinement – all at petabyte scale.
Hadoop certainly brings tremendous value in its ability to handle massive and diverse amounts of data, but the sheer complexity makes Hadoop challenging to all but the most seasoned (and scarce) technologists and data scientists. For the subset of companies out there whose core operations involve data engineering, this may not be a problem. But, a large number of businesses remain unable to exploit big data for competitive advantage, and say that integrating technologies is an obstacle to deployments.
That’s why I was proud as part of my keynote address this week in San Jose at the sixth annual Hadoop Summit to announce the release of Teradata Portfolio for Hadoop, a comprehensive offering that takes the complexity and cost out of deploying and managing Hadoop. Simply stated, Teradata Portfolio for Hadoop is a single source for all things Hadoop that removes the engineering and operations guesswork for non-programmers, business analysts and business consumers looking to extract value and competitive advantage from their data in an unprecedented range of commercial environments.
You can learn a lot more in our press release, but even a quick summary makes the benefits clear. Backed by our partner Hortonworks, Teradata Portfolio for Hadoop includes flexible Apache Hadoop product platforms and software, along with best-in-class services, training and customer support and direct access to the latest technologies coming from the open source Apache Hadoop community.
Every organization is different, so Teradata Portfolio for Hadoop offers four implementation choices. If your IT department wants turnkey, ready-to-run solutions, we deliver two premium platforms: Teradata Appliance for Hadoop, a tightly integrated hardware and software appliance optimized for enterprise-class data storage and management; and Teradata Aster Big Analytics Appliance, for data staging and processing with more than 70 pre-packaged analytic functions and optional nodes of Hadoop. For engineering-led departments that opt to build their own systems but want the backing of Teradata, we’ve developed two commodity platforms for Hadoop: Teradata Software Only for Hadoop and Teradata Commodity Offering for Hadoop, the latter being a partnership with Dell for a lower cost, standard hardware configuration optimized for the Hortonworks Data Platform.
This variety of Hadoop implementations gives organizations real choices depending on the level of enterprise expertise and budgetary constraints. And Teradata, and Hortonworks, offers additional assistance with software, consulting services and customer support. We can help with installation, mentoring and knowledge transfer, operations, management, administration, backup, security processes and more to ensure end-to-end analytical solutions for enterprises needing to manage most any type of data — be it social media, documents, text, images, audio or any variety of multi-structured data.
Exponential data growth and velocity demand an ever-faster ability to capture, store and refine data. I’m proud to say our joint solutions with Hortonworks expands these capabilities beyond the rarified world of companies that can afford to hire a corps of data scientists commanding high annual salaries. “Main street” organizations now have comprehensive, intuitive and flexible options to leverage Hadoop’s tremendous power in all sorts of innovative ways.