Getting Started With Apache Hadoop on CentOS Linux
Ubuntu Hadoop Quick Start Guides
The Linked Tutorials are Showing How to Install and Getting-Started with Apache Hadoop/Map-Reduce for Big Data on CentOS Linux Distribution.
Hadoop is an open source framework for writing and running Distributed
Applications that Process Big Data (large amounts of data).
Apache Hadoop key Features:
- Accessible — Hadoop runs on large clusters of commodity machines or on cloud computing services such as Amazon’s Elastic Compute Cloud (EC2).
- Robust — Because it is intended to run on commodity hardware, Hadoop is architected with the assumption of frequent hardware malfunctions. It can gracefully handle most such failures.
- Scalable — Hadoop scales linearly to handle larger data by adding more nodes to the cluster.
- Simple — Hadoop allows users to quickly write efficient parallel code.