How to Install & Getting Started With Apache Hadoop Database for Linux Kubuntu 13.04 Raring Desktop

May 22, 2013 | By Duchateaux.

Install Hadoop for Kubuntu 13.04 Raring 32/64-bit

Hi! The Tutorial shows you Step-by-Step How to Install and Getting-Started with Apache Hadoop/Map-Reduce vanilla in Pseudo-Distributed mode on Linux Kubuntu 13.04 Raring i386/amd64 Desktop.

The Guide Describe a System-Wide Setup with Root Privileges but you Can Easily Convert the Procedure to a Local One.

The Contents and Details are Expressly Essentials to Give Focus Only to the Essentials Instructions and Commands.

Install Hadoop for Kubuntu 13.04 Raring - Featured

  1. Download Latest Apache Hadoop Stable Release:

    Apache Hadoop Binary tar.gz
  2. Right-Click on Archive > Open with Ark

    Then Extract Into /tmp.

    KDE4 Apache Hadoop Stable tar.gz Extraction
  3. Open Terminal Window
    (Press “Enter” to Execute Commands)

    Linux Kubuntu Open Terminal

  4. Relocate Apache Hadoop Directory

    sudo su

    If Got “User is Not in Sudoers file” then see: How to Enable sudo

    mv /tmp/hadoop* /usr/local/
    ln -s /usr/local/hadoop* /usr/local/hadoop
    mkdir /usr/local/hadoop/tmp
    sudo chown -R root:root /usr/local/hadoop*

  5. How to Install Oracle Official Java JDK on Kubuntu

  6. Set JAVA_HOME in Hadoop Env File

    nano /usr/local/hadoop/conf/hadoop-env.sh

    Inserts:

    /nexport JAVA_HOME=/usr/lib/jvm/<oracleJdkVersion>

    Ctrl+x to Save & Exit :)

  7. Configuration for Pseudo-Distributed mode

    nano /usr/local/hadoop/conf/core-site.xml

    The Content Should Look Like:

    <?xml version=”1.0″?>
    <?xml-stylesheet type=”text/xsl” href=”configuration.xsl”?>
    <configuration>
    <property>
    <name>hadoop.tmp.dir</name>
    <value>/usr/local/hadoop/tmp</value>
    </property>
    <property>
    <name>fs.default.name</name>
    <value>hdfs://localhost:8020</value>
    </property>
    </configuration>

    Next:

    nano /usr/local/hadoop/conf/hdfs-site.xml

    The Content Should Look Like:

    <?xml version=”1.0″?>
    <?xml-stylesheet type=”text/xsl” href=”configuration.xsl”?>
    <configuration>
    <property>
    <name>dfs.replication</name>
    <value>1</value>
    </property>
    <property>
    <!– specify this so that running ‘hdfs namenode -format’
    formats the right dir –>
    <name>dfs.name.dir</name>
    <value>/usr/local/hadoop/cache/hadoop/dfs/name</value>
    </property>
    </configuration>

    Latest:

    nano /usr/local/hadoop/conf/mapred-site.xml

    The Content Should Look Like:

    <?xml version=”1.0″?>
    <?xml-stylesheet type=”text/xsl” href=”configuration.xsl”?>
    <configuration>
    <property>
    <name>mapred.job.tracker</name>
    <value>localhost:8021</value>
    </property>
    </configuration>

  8. SetUp Path & Environment

    su <myuser>
    cd  
    nano .bashrc

    Inserts:

    HADOOP_HOME=/usr/local/hadoop
    /nexport PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin

    The JAVA_HOME is Set Following Oracle Java JDK6+ Installation Version…

    Then Load New Setup:

    source $HOME/.bashrc
  9. SetUp Needed Local SSH Connection

    sudo systemctl start ssh

    Generate SSH Keys to Access:

    ssh-keygen -b 2048 -t rsa
    echo "$(cat ~/.ssh/id_rsa.pub)" > ~/.ssh/authorized_keys

    Testing Connection:

    ssh 127.0.0.1
  10. Formatting HDFS

    hdfs namenode -format

    Terminal Apache Hadoop HDFS Formatting Success

  11. Starting Up Hadoop Database

    start-all.sh

  12. Apache Hadoop Database Quick Start Guide


QuickChic Theme • Powered by WordPress