Introduction
In the previous chapter, we discussed how to write MapReduce programs in various ways in order to analyze data. Earlier, MapReduce was the only means of processing data in Hadoop, but with the passage of time, growing popularity, and the ease of using the Hadoop platform, various subprojects joined its league, which helps users write logic in SQL, scripts, and so on, and analyze the data. Projects, such as Apache Hive, Apache Pig, and Apache Hbase, are well accepted by users, so the majority of development these days takes place using any of these. In this chapter, we are going to take a look at how to use these tools to perform various advanced operations. If you are new to these terms, I would recommend that you read through these links first.
For Hive resources, take a look a these links:
http://hadooptutorials.co.in/tutorials/hive/introduction-to-apache-hive.html
http://hadooptutorials.co.in/tutorials/hive/hive-data-units.html