Hortonworks and Teradata believe that no single analytic system can meet all customers’ needs, and leading data driven organizations will deploy an analytical ecosystem. Harnessing relational and non-relational data, and running SQL and non-SQL analytics requires specialized technologies. That is why Hortonworks and Teradata have a longstanding engineering partnership focused on delivering the analytical ecosystem to the enterprise through best in class analytic engines and co-engineering of software to orchestrate the processing and strip out the complexity typically associated with a hybrid environment.
UNIFIED DATA ARCHITECTURE
As organizations embrace the idea of an ecosystem approach that leverages Hadoop alongside relational and other environments to harness all data, the challenge becomes connecting the dots data stored in different environments.
Cross-Platform Access to All Your Data and Analytic Engines
Let your business work with a seamless data fabric across all of your data and analytical engines by executing a single query that pushes down processing into multiple analytics engines in parallel with Teradata QueryGrid™. Users get the most value out of all your data by taking advantage of specialized processing engines operating as a cohesive analytic environment.
A Comprehensive Ingest Framework
An intelligent, self-service software solution for ingesting and distributing extremely fast moving data streams throughout the analytical ecosystem is possible with Teradata Listener™ and Kylo. Teradata Listener’s easy to use interface, built in intelligence and tight integrations with HDF, makes it the idea fit for business users and DevOps teams who need to stream new data sets into any location in the ecosystem.
Kylo is a next generation, open source software with subscription support from Think Big for creating a Data Lake solution:
Kylo has been built on eight years of global expertise involving 150 data lake projects in global banking, telecoms, retail and other industries. Kylo is currently in private beta. It will be available with an Apache open source license in Q1 2017.
Faster Time to Value with Hadoop
Organizations that need a powerful, ready to-run enterprise platform that is pre-configured and optimized specifically for Hadoop turn to the Teradata® Appliance for Hadoop. As a purpose-built, integrated hardware and software solution for data at scale, the appliance runs the latest version of HDP with the additional benefits of simplified manageability and enterprise grade reliability that comes from an engineered appliance, all backed by world class customer support from Teradata. By minimizing the number of moving parts required for deployment and operations, the appliance allows companies to achieve faster time to value by just plugging the appliance into existing infrastructure, thereby leveraging your current investments in technology and resources.
The World’s Most Powerful Database
The market-leading platform for delivering strategic and operational analytics throughout your organization so users access a single source of consistent, centralized, integrated data is available with the Teradata Database. Teradata’s approach to integrated data drives the highest business value through cross-functional analysis. Learn more about the Teradata Database and Integrated Data Warehouse.
Integrated Big Data Analytics
Organizations attain unmatched competitive advantage and drive pervasive adoption of big data analytics using Aster Analytics on Hadoop. The platform features Teradata Aster SQL-GR™ analytic engine, a native graph processing engine for graph analysis across big data sets, SQL and SQL-MapReduce® analytic engines that enable a variety of analytics like SQL analysis, path/pattern analysis, statistical analysis and text analysis. Learn more about Teradata Aster Big Data Analytics.
Aster Analytics on Hadoop offers more than 100 prebuilt, fully parallel analytic functions that are callable via SQL or R, popular languages for business analytics. The prebuilt functions eliminate the need for organizations to create or install their own, which accelerates time to value. Since these functions are accessible by SQL, analysts can easily combine different analytics into a single work stream.
Aster Analytics on Hadoop runs in existing Hortonworks Enterprise Hadoop clusters, so there’s no additional footprint. Aster processing engines are easily provisioned and managed by YARN, allowing users to virtualize multiple instances of Aster Analytics within a single Hadoop cluster to support analytics for different departments, data science labs and projects. The solution can easily access data in the Hadoop Distributed File System (HDFS) so users can explore, discover and analyze data where it resides. No data movement across systems is required. In addition, no data creation in an Aster format is needed. Users also have the unique capability to provision an exploration and development sandbox or production system in Hadoop. Once the model is complete, it can be deployed into a production instance in the same Hadoop cluster, accessing the same data.
A Powerful SQL Engine for Hadoop and Beyond
Originally developed by Facebook, Presto is an open source distributed SQL query engine for running interactive analytic queries against data sources of all sizes ranging from gigabytes to petabytes. Presto’s pure memory-based architecture is built for speed, allowing Presto to support large numbers of concurrent interactive queries against huge data sets. Building upon the Presto query engine developed and open sourced by Facebook, Teradata has made a multiyear commitment to contribute to Presto’s open source development and provide commercial support designed to increase adoption of this free, 100% open source software. Learn more about Presto or download Presto now.
Think Big provides expert advisory and implementation services for open-source big data solutions. As the first and only pure-play big data services firm, our data scientists and engineers are trusted advisors to the world’s most innovative companies. Our experienced teams combine a distinctive methodology and a proven Think Big Velocity framework that includes tested design patterns and pre-built components to help clients build applications faster.
We work with our customers every step of the way, identifying individual goals and collaborating to achieve effective and growth driven solutions. Using holistic thinking and proactive planning, our experts will guide you through every step of the big data lifecycle to deliver value quickly and scale seamlessly.
We are the leaders in data lake implementations – with dozens of successful engagements, a full set of technology best practices, and pre-built solution framework components to work with a variety of platforms and tools that accelerate time-to-value.
Our skilled architects and engineers design and build big data solutions that produce faster time-to-value, with clear architectural blueprints for the long term. We build on a foundation of reuse for frameworks and components that enable companies to unlock the value of their data much sooner than other methods.
We help companies build new business capabilities from running analytics on big data technologies, such as Spark and Hadoop and go beyond just helping you use various data science libraries and other tools, turning new ideas and opportunities into scalable and production-ready solutions.
We offer robust and comprehensive managed services to support big data platforms and applications. Our experienced big data experts use advanced tools and well-defined, field-tested processes based on ITIL standards to deliver continuous improvements for Hadoop platforms and big data environments.
Think Big Academy helps develop advanced employee skills to utilize big data for better business decisions. Our certified experts provide a variety of courses in Apache Hadoop, Apache Spark, Apache Cassandra and Big Data Concepts.
Teradata Corporation (NYSE: TDC) is a global leader in analytic data platforms and consulting services. Teradata helps organizations collect, integrate, and analyze all of their data so they can know more about their customers and business and do more of what’s really important. Visit teradata.com for details.
HDP - HDP Certified badge indicates this partner’s solution has been certified to work with HDP; reviewed for architectural best practices and validated against a comprehensive suite of integration test cases, benchmarked for scale under varied workloads and comprehensively documented.
Yarn Ready - Apache Hadoop YARN is the data operating system for Hadoop 2. YARN Ready certification recognizes applications that integrate with YARN and process data via pushdown computation to the cluster. Examples of a YARN ready solution includes an application that has native YARN application master or leverages scale-out capabilities of the platform like Hive, Spark and MR2.
Operations Ready - Apache Ambari is the open source management fabric for Hadoop. Operations Ready certification recognizes applications that can either be deployed or managed via Ambari or integrate with Ambari's APIs.