Apache Hadoop modules
Apache Hadoop consists of the following modules:
Hadoop Common: This is a common module used by other modules of Hadoop. It is like a utility package.
Hadoop Distributed File System: Hadoop Distributed File System can be considered when we have to store large amounts of data across various machines or machine clusters.
Hadoop Yarn: Think of a scenario where we have many servers on the cloud that need to be scheduled to restart or reboot at a particular time by sending an e-mail intimation to the tenants. Hadoop Yarn can be used for scheduling resources across computers or clusters.
Hadoop Map and Reduce: If we have to process a large set of data, we can break it into small clusters and process them as units and merge them back later. This can be done with the libraries provided in Apache map and reduce.
Spring namespace for Hadoop
Following is the namespace that needs to be used to integrate the Hadoop framework with Spring. http://www.springframework.org/schema/hadoop...