Translating your SQL Skills to Hadoop
Hadoop jobs have grown 200,000%. No, that’s not a typo. According to Indeed.com, Hadoop is one of the top 10 job trends right now.
When you look at LinkedIn, the growth in profiles that have SQL in them is on the downswing — about -4%, but the growth of profiles that have Hadoop in them is up 37%. Hadoop is becoming a clear resume differentiator. Updating and maintaining technical skills has always been part of the job and is part of ensuring a long and healthy career. But that can be easier said than done.
Translating SQL Skills to Hadoop
The need for RDBMS (and other DB) skills is not going away, and Hadoop isn’t a replacement for those systems. Rather it’s an augmentation of those systems for scenarios involving the volumes of data that they can’t efficiently tackle. There’s still a robust market tons of job listings, but not a lot of growth.
Does learning Hadoop mean you have to start all over again? No, learning Hadoop is an evolution, not a revolution. While Hadoop may be a hot new technology, if you’re already comfortable with SQL, then learning how to use Hive (one of the key projects within the Hadoop ecosystem) will feel very familiar as it’s syntax and commands are very SQL-like – deliberately so.
Hortonworks’ Stinger Initiative has the goals of improving the performance of Hive to familiar ‘interactive time’ querying across large data sets and moving ever closer to full compliancy with SQL-92. If you’re already comfortable with SQL, then this technology shift will feel less like writing with the other hand but more like learning to write with a new pen. It’s not a revolution in your skills, it’s simply an evolution.
Not only is Hive very SQL-like, Hadoop is a “+1″ in your environment, a perfect complement to your existing data warehouse, storage and database infrastructures. As part of the modern data architecture, Hadoop enables data refinement and exploration at scale in a cost effective way across commodity hardware and unlocks the value of your big data to enrich your apps and experiences. With Hadoop as an addition to your architecture, you can continue to run the infrastructure you’re comfortable with and yet still gain all of the advantages of Hadoop.
So, what’s the easiest way for you to learn more about Hive? We have a simple 3 step process for you:
- Read about Hadoop or watch the Hadoop Tutorial: Core Apache Hadoop
- Read about Hive or watch the Hadoop Tutorial: Apache Hive
- You’ll find lots of other information on our website and videos on our YouTube channel — if you’re new to Hadoop, you can see Hadoop in action in these Video Tutorials: Using Hadoop.
2) Download the Hortonworks Sandbox
It’s a free, single-node, personal Hadoop environment that is a learning platform as well. You can follow along with the tutorials in the Sandbox, and then safely go off-roading to experiment and learn as the Sandbox is a complete Hortonworks Data Platform environment where you can upload your own datasets and connect to your favorite BI tools. No AWS accounts needed, no space in your data center required, allowing you to run a fully functional node of Hadoop on your laptop while you’re in a plane, train or automobile.
3) Attend a Hortonworks University Class
- New to Hadoop or want a business-level overview? We have a one-day, live, online Hadoop Essentials course
- Are you comfortable developing in Java, take a look at the Developing Apache Hadoop Solutions in Java
- Do you develop apps in a Windows server environment, then look at Developing Solutions for Hadoop on Windows
- Are you a Sys Admin or a DBA and want to know how to administer a Hadoop Cluster? Look at the Hadoop for Administrators
- Are you a Data Scientist? Look at the Applying Data Science Using Apache Hadoop
*Note, these statistics are valid as of July 2013.
Try it with Sandbox
Hortonworks Sandbox is a self-contained virtual machine with Apache Hadoop pre-configured alongside a set of hands-on, step-by-step Hadoop tutorials.