Get fresh updates from Hortonworks by email

Once a month, receive latest insights, trends, analytics information and knowledge of Big Data.


Sign up for the Developers Newsletter

Once a month, receive latest insights, trends, analytics information and knowledge of Big Data.


Get Started


Ready to Get Started?

Download sandbox

How can we help you?

* I understand I can unsubscribe at any time. I also acknowledge the additional information found in Hortonworks Privacy Policy.
closeClose button
May 07, 2014
prev slideNext slide

DMX-h from Syncsort now Certified on Hortonworks Data Platform 2.1

Syncsort is a Hortonworks Certified Technology Partner and has over 40 years of experience helping organizations integrate big data…smarter. Keith Kohl, Director of Product Management, Syncsort, is our guest blogger. Below he talks about the importance of certification and how it benefits Syncsort’s customers and prospects interested in Hadoop.

Back in January, Syncsort announced our partnership with Hortonworks and the certification of DMX-h on HDP 2.0. I was also given the opportunity to write a guest BLOG on the Hortonworks site about HDP 2 and the GA of YARN (thanks Hortonworks!).

Syncsort integrates with Hadoop and HDP directly through YARN, making it easier for users to write and maintain MapReduce jobs graphically. If you want to see, there are a number of videos here.

The Hadoop ecosystem is moving quickly, and so are we here at Syncsort. With the new release of HDP 2.1, we are excited to announce the certification of DMX-h on HDP 2.1!

Syncsort and Hortonworks in the Modern Data Architecture

This means our mutual customers have the confidence that the products are integrated and work together out of the box. Additionally, through the YARN integration, the processing initiated by DMX-h within the HDP cluster will make better use of the resources and execute more efficiently.  This certification reflects our ongoing commitment to continue to work with Hortonworks and contribute to Apache Hadoop projects for the benefit of the entire Big Data community. As I mentioned in my previous BLOG, ETL is a common use case for Hadoop, even if users don’t know they’re doing ETL—they could be calling it data refinement, data preparation, data management or something else. But ETL is the most common use case.

But the question is, weren’t these organizations doing ETL before?  So, why are they switching to Hadoop? The answer is to move processing from platforms such as the data warehouse (doing ELT), to a cost effective environment such as Hadoop. I’ve heard this called data warehouse optimization or offload. This BLOG isn’t about offload, but you can read more about it here from us, and from Hortonworks. And you’re going to hear a lot more about offload from us…stay tuned!

Hortonworks did something pretty cool by providing users with a VM of a completely installed version of HDP called the Hortonnworks Sandbox.  We took that Sandbox and gave users the ability to download DMX-h and install it on the Hortonworks Sandbox.  We also include some sample job templates – like the use cases above – and sample data.

You can see for yourself here, Syncsort – Hortonworks Sandbox Tutorial.

Keith Kohl, Dir. Product Management, Syncsort



Leave a Reply

Your email address will not be published. Required fields are marked *