1. Create a new user 'orienit' and folder '/home/orienit/spark' to download and extract Hadoop files.
2. Configure environment variables in ~/.bashrc file and generate SSH keys for secure communication.
3. Update configuration files in $HADOOP_HOME/etc and /etc/hosts file to set up the single node Hadoop cluster.
4. Format and start the Hadoop daemons using start-all.sh and stop using stop-all.sh.
1. Create a new user 'orienit' and folder '/home/orienit/spark' to download and extract Hadoop files.
2. Configure environment variables in ~/.bashrc file and generate SSH keys for secure communication.
3. Update configuration files in $HADOOP_HOME/etc and /etc/hosts file to set up the single node Hadoop cluster.
4. Format and start the Hadoop daemons using start-all.sh and stop using stop-all.sh.
1. Create a new user 'orienit' and folder '/home/orienit/spark' to download and extract Hadoop files.
2. Configure environment variables in ~/.bashrc file and generate SSH keys for secure communication.
3. Update configuration files in $HADOOP_HOME/etc and /etc/hosts file to set up the single node Hadoop cluster.
4. Format and start the Hadoop daemons using start-all.sh and stop using stop-all.sh.
-------------------------------------------------------------------------------------------------------Hadoop Single Node setup
-------------------------------------------------------------------------------------------------------1. create a new 'orienit' user in ubuntu.
2. create '/home/orienit/spark' folder mkdir /home/orienit/spark 3. Download 'hadoop-2.6.0.tar.gz' file from hadoop repository, copy 'hadoop-2.6. 0.tar.gz' file into this '/home/orienit/spark' directory and extract the tar fil e in same directory tar -xvzf hadoop-2.6.0.tar.gz . 4. update '~/.bashrc' file with below changes command: gedit ~/.bashrc export JAVA_HOME=/usr/lib/jvm/java-1.7.0-openjdk export HADOOP_HOME=/home/orienit/spark/hadoop-2.6.0 export PATH=$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH 5. Enter the below commands on terminal: ssh localhost ssh-keygen -t rsa -P '' -f ~/.ssh/id_rsa cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys 6. Update the '$HADOOP_HOME/etc/hadoop' folder files 'core-site.xml','hdfs-site. xml','mapred-site.xml','yarn-site.xml','hadoop-env.sh','mapred-env.sh','yarn-env .sh' and 'slaves' files as per the required configurations 7. update "/etc/hosts" file. command: sudo gedit /etc/hosts 127.0.0.1 127.0.0.1
localhost hadoop
8. Format the 'namenode' from current machine using this command:
hadoop namenode -format 9. Start the hadoop by using this command on current machine: start-all.sh mr-jobhistory-daemon.sh start historyserver 10. Stop the hadoop by using this command on current machine: stop-all.sh mr-jobhistory-daemon.sh stop historyserver --------------------------------------sudo su <enter ur password> cp -r /home/<src_username>/Desktop/hadoop_2.x_installation /home/<dest_username> /Desktop/
x_installation exit example: copy data from 'orienit' user to 'demo' user -------------------------------------------------sudo su <enter ur password> cp -r /home/orienit/Desktop/hadoop_2.x_installation /home/demo/Desktop/ chown -R demo:demo /home/demo/Desktop/hadoop_2.x_installation exit --------------------------------------change hostname