$schemamarkup = get_post_meta(get_the_ID(), 'schemamarkup', true); if(!empty($schemamarkup)) { echo $schemamarkup; }

How to Install & Getting Started With Apache Hadoop Database for Linux Ubuntu 13.04 Raring Desktop

May 22, 2013 | By Duchateaux.

Hi! The Tutorial shows you Step-by-Step How to Install and Getting-Started with Apache Hadoop/Map-Reduce vanilla in Pseudo-Distributed mode on Ubuntu Linux 13.04 Raring i386/amd64 Desktop.

The Guide Describe a System-Wide Setup with Root Privileges but you Can Easily Convert the Procedure to a Local One.

The Contents and Details are Expressly Essentials to Give Focus Only to the Essentials Instructions and Commands.

Gnome Penguin Hadoop

  1. Download Latest Apache Hadoop Stable Release:

    Apache Hadoop Binary tar.gz
  2. Double-Click on Archive and Extract Into /tmp Directory

    Ubuntu Unity Extract Apache Hadoop Stable tar.gz Archive

  3. Open Terminal Window
    (Press “Enter” to Execute Commands)

    Cmd and Search “term”

    Ubuntu Linux Raring Open Terminal

    In case first see: Terminal QuickStart Guide.

  4. Relocate Apache Hadoop Directory

    sudo su

    If Got “User is Not in Sudoers file” then see: How to Enable sudo

    mv /tmp/hadoop* /usr/local/
    ln -s /usr/local/hadoop* /usr/local/hadoop
    mkdir /usr/local/hadoop/tmp
    sudo chown -R root:root /usr/local/hadoop*

  5. How to Install Oracle Official Java JDK on Ubuntu

    Install Oracle JDK for Ubuntu
  6. Set JAVA_HOME in Hadoop Env File

    nano /usr/local/hadoop/conf/hadoop-env.sh

    Inserts:

    /nexport JAVA_HOME=/usr/lib/jvm/<oracleJdkVersion>

    Ctrl+x to Save & Exit :)

  7. Configuration for Pseudo-Distributed mode

    nano /usr/local/hadoop/conf/core-site.xml

    The Content Should Look Like:

    <?xml version=”1.0″?>
    <?xml-stylesheet type=”text/xsl” href=”configuration.xsl”?>
    <configuration>
    <property>
    <name>hadoop.tmp.dir</name>
    <value>/usr/local/hadoop/tmp</value>
    </property>
    <property>
    <name>fs.default.name</name>
    <value>hdfs://localhost:8020</value>
    </property>
    </configuration>

    Next:

    nano /usr/local/hadoop/conf/hdfs-site.xml

    The Content Should Look Like:

    <?xml version=”1.0″?>
    <?xml-stylesheet type=”text/xsl” href=”configuration.xsl”?>
    <configuration>
    <property>
    <name>dfs.replication</name>
    <value>1</value>
    </property>
    <property>
    <!– specify this so that running ‘hdfs namenode -format’
    formats the right dir –>
    <name>dfs.name.dir</name>
    <value>/usr/local/hadoop/cache/hadoop/dfs/name</value>
    </property>
    </configuration>

    Latest:

    nano /usr/local/hadoop/conf/mapred-site.xml

    The Content Should Look Like:

    <?xml version=”1.0″?>
    <?xml-stylesheet type=”text/xsl” href=”configuration.xsl”?>
    <configuration>
    <property>
    <name>mapred.job.tracker</name>
    <value>localhost:8021</value>
    </property>
    </configuration>

  8. SetUp Path & Environment

    su <myuser>
    cd  
    nano .bashrc

    Inserts:

    HADOOP_HOME=/usr/local/hadoop
    /nexport PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin

    The JAVA_HOME is Set Following Oracle Java JDK6+ Installation Version…

    Then Load New Setup:

    source $HOME/.bashrc
  9. SetUp Needed Local SSH Connection

    sudo systemctl start ssh

    Generate SSH Keys to Access:

    ssh-keygen -b 2048 -t rsa
    echo "$(cat ~/.ssh/id_rsa.pub)" > ~/.ssh/authorized_keys

    Testing Connection:

    ssh 127.0.0.1
  10. Formatting HDFS

    hdfs namenode -format

    Terminal Apache Hadoop HDFS Formatting Success

  11. Starting Up Hadoop Database

    start-all.sh

  12. Apache Hadoop Database Quick Start Guide

    Hadoop MapReduce Quick Start