4. If you are using putty to access your Linux
box remotely, please install openssh by
running this command, this also helps in
configuring SSH access easily in the later
part of the installation:
sudo apt-get install openssh-server
8. 2. Adding dedicated Hadoop
system user.
a. Adding group:
sudo addgroup hadoop
b. Creating a user and adding the user to
a group:
sudo adduser –ingroup hadoop hduser
13. i. Run this following command to download
Hadoop version 2.2.0
wget http://mirrors.cnnic.cn/apache/hadoop/common/hadoop-
2.2.0/hadoop-2.2.0.tar.gz
ii. Unpack the compressed hadoop file by using
this command:
tar -xvzf hadoop-2.2.0.tar.gz
iii. move hadoop-2.2.0 to hadoop directory by
using give command
mv hadoop-2.2.0 hadoop
14. iv. Move hadoop package of your choice
sudo mv hadoop /usr/local/
v. Make sure to change the owner of all the files
to the hduser user and hadoop group by using
this command:
cd /usr/local/
sudo chown -R hduser:hadoop hadoop
16. The following are the required files we will use
for the perfect configuration of the single
node Hadoop cluster.
a. yarn-site.xml:
b. core-site.xml
c. mapred-site.xml
d. hdfs-site.xml
e. Update $HOME/.bashrc
We can find the list of files in Hadoop
directory which is located in
cd /usr/local/hadoop/etc/hadoop
25. i. The first step to starting up your Hadoop
installation is formatting the Hadoop
filesystem which is implemented on top of
the local filesystem of your cluster. You
need to do this the first time you set up a
Hadoop cluster. Do not format a running
Hadoop filesystem as you will lose all the
data currently in the cluster (in HDFS).
hadoop namenode -format
26. ii. Start Hadoop Daemons by running the
following commands:
Name node:
hadoop-daemon.sh start namenode
Data node:
hadoop-daemon.sh start datanode