1) (i) Perform setting up and Installing Hadoop in its three operating modes:
a) Standalone
b) Pseudo distributed
c) Fully distributed
(ii) Use web based tools to monitor your Hadoop setup. Installation of Standalone mode
2) Implement the following file management tasks in Hadoop:
Adding files and directories.
Retrieving files.
Deleting files
Hint: A typical Hadoop workflow creates data files (such as log files) elsewhere and copies them into HDFS using one of the above command line utilities.
3) (i) Run a basic Word Count Map Reduce program to understand Map Reduce Paradigm.(Old API)
(ii) Run a basic Word Count Map Reduce program to understand Map Reduce Paradigm.(New API)
4) Write a Map Reduce program that mines weather data.
5) (i) Perform setting up and Installing SPARK.
6) Run a basic Word Count Map Reduce program in scala using SPARK
7) Write a SPARK program that mines weather data. Weather sensors collecting data every hour at many locations across the globe gather a large volume of log data, which is a good candidate for analysis, since it is semi structured and record-oriented.
8) Install and Run Pig then write Pig Latin scripts to sort, group, join, project, and filter your data.
9) Install and Run Hive then use Hive to create, alter, and drop databases, tables, views, functions, and indexes.