Professional Documents
Culture Documents
BAI-7
Version 1.0
Abstract
This is the lab guide for the participants to learn and complete all lab exercises as part of the
Apache Spark Developer training.
Manaranjan Pradhan
manaranjan@enablecloud.com
https://www.linkedin.com/in/manaranjanpradhan
Apache Spark Developer Training – Lab Guide
Manaranjan Pradhan
Page |1
manaranjan@enablecloud.com
https://www.linkedin.com/in/manaranjanpradhan
Apache Spark Developer Training – Lab Guide
Manaranjan Pradhan
Page |2
manaranjan@enablecloud.com
https://www.linkedin.com/in/manaranjanpradhan
Apache Spark Developer Training – Lab Guide
o Enter Password
▪ Password is hadoop123
• Click “Continue”
Manaranjan Pradhan
Page |3
manaranjan@enablecloud.com
https://www.linkedin.com/in/manaranjanpradhan
Apache Spark Developer Training – Lab Guide
• Finally select the VM on the oracle virtual box console and click on the START button
• Select “Hadoop Developer” user
o Enter Password
▪ Password is hadoop123
3) Configuring the VM
a. Get your VM’s IP address
Right click on your VM console and select ‘open in terminal’
Manaranjan Pradhan
Page |4
manaranjan@enablecloud.com
https://www.linkedin.com/in/manaranjanpradhan
Apache Spark Developer Training – Lab Guide
Change the following ip address to the one obtained in the previous step
Save and exit the file. And verify if the settings are working file by running the following
command.
ping hadooplab.bigdataleap.com
If you are getting reply from the VM, then it is configured properly.
Note: MAC users should update their /etc/hosts file to add VM’s hostname and IP address
Save and exit the file. And verify if the settings are working file by running the following
command in the windows command shell.
ping hadooplab.bigdataleap.com
If you are getting reply from the VM, then it is configured properly.
cd /home/hadoop/lab
Manaranjan Pradhan
Page |5
manaranjan@enablecloud.com
https://www.linkedin.com/in/manaranjanpradhan
Apache Spark Developer Training – Lab Guide
/home/hadoop/lab contains the following directories and will be used for the following
purposes.
lab/software – hadoop, spark and hive is installed in this software. This is the home
directory for all software.
lab/cluster – configured for hadoop internal directories to store blocks, fsimages and
temporary working file for the framework.
Manaranjan Pradhan
Page |6
manaranjan@enablecloud.com
https://www.linkedin.com/in/manaranjanpradhan