Skip to content

Hadoop Workflow Example

Hadoop is a scalable, distributed computing solution provided by Apache. Similar to queuing systems, Hadoop allows for distributed processing of large data sets.

Installing & Running Hadoop

Note

The flight environment will need to be activated before the environments can be created so be sure to run flight start or setup your environment to automatically activate the flight environment.

  • Install dependencies for Hadoop:
    [flight@chead1 (mycluster1) ~]$ sudo yum install -y java-1.8.0-openjdk.x86_64 java-1.8.0-openjdk-devel.x86_64
    
  • Download Hadoop v3.2.1:

    [flight@chead1 (mycluster1) ~]$ flight silo software pull --repo openflight hadoop 3.2.1
    

    Tip

    If you are using a version of java other than 1.8.0 then the version will need to be changed on line 54 in SILO_SOFTWARE_DIR/hadoop/3.2.1/etc/hadoop/hadoop-env.sh

  • Add the hadoop installation to the user's path along with the Java home (replacing SILO_SOFTWARE_DIR with the software directory used by silo in the download above, this can be done temporarily in the CLI or by adding to the user's ~/.bashrc):

    export PATH="$PATH:SILO_SOFTWARE_DIR/hadoop/3.2.1/bin/:SILO_SOFTWARE_DIR/hadoop/3.2.1/sbin/"
    export JAVA_HOME="/usr/lib/jvm/java-1.8.0/jre"
    

    Note

    If the above has been set in the ~/.bashrc then a new session for the user will need to be started for the environment changes to take effect, otherwise the below commands will not be located

  • Start the Hadoop distributed file system service:

    [flight@chead1 (mycluster1) MapReduceTutorial]$ start-dfs.sh
    

  • Start the resource manager, node manager and app manager service:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ start-yarn.sh
    

Downloading the Hadoop Job

These steps help setup the Hadoop environment and download a spreadsheet of data which will Hadoop will sort into sales units per region.

  • Download and source Hadoop environment variables:
    [flight@chead1 (mycluster1) ~]$ flight silo file pull openflight:hadoop/hadoopenv
    [flight@chead1 (mycluster1) ~]$ source hadoopenv
    

Tip

Be sure to update line 1 in hadoopenv to match the installation location of hadoop. If using a different version of java, update line 3.

  • Create job directory:
    [flight@chead1 (mycluster1) ~]$ mkdir MapReduceTutorial
    [flight@chead1 (mycluster1) ~]$ chmod 777 MapReduceTutorial
    
  • Download job data:
    [flight@chead1 (mycluster1) ~]$ cd MapReduceTutorial
    [flight@chead1 (mycluster1) MapReduceTutorial]$ flight silo file pull openflight:hadoop/hdfiles.zip
    [flight@chead1 (mycluster1) MapReduceTutorial]$ unzip -j hdfiles.zip
    
  • Check that job data files are present:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ ls
    desktop.ini  hdfiles.zip  SalesCountryDriver.java  SalesCountryReducer.java  SalesJan2009.csv  SalesMapper.java
    

Preparing the Hadoop Job

  • Compile java for job:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ javac -d . SalesMapper.java SalesCountryReducer.java SalesCountryDriver.java
    
  • Create a manifest file:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ echo "Main-Class: SalesCountry.SalesCountryDriver" >> Manifest.txt
    
  • Compile the final java file for job:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ jar cfm ProductSalePerCountry.jar Manifest.txt SalesCountry/*.class
    

Loading Data into Hadoop

  • Create directory for processing data and copy sales results in:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ mkdir ~/inputMapReduce
    [flight@chead1 (mycluster1) MapReduceTutorial]$ cp SalesJan2009.csv ~/inputMapReduce/
    
  • Load the data into the distributed file system:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ $HADOOP_HOME/bin/hdfs dfs -ls ~/inputMapReduce
    

Running the Hadoop Job

  • Execute the MapReduce job:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ $HADOOP_HOME/bin/hadoop jar ProductSalePerCountry.jar ~/inputMapReduce ~/mapreduce_output_sales
    
  • View the job results:
    [flight@chead1 (mycluster1) MapReduceTutorial]$ $HADOOP_HOME/bin/hdfs dfs -cat ~/mapreduce_output_sales/part-00000 | more