Setup Of Hadoop Cluster ; Uploading, Reading, Deleting Files Through Cluster.
Shivani Agarwal
Software Developer @ThinkTech Software Inc. | Full Stack | React | Python | DSA
Hadoop Cluster is use to solve BigData problems. Hadoop is product of Apache and it is very popular in BigData world, facebook is using this to solve there BigData problem to Store there data of user.
So let's see how to build Hadoop cluster to solve bigdata problem
To create this hadoop cluster we need some basic things first;
- O.S. ( we are using RedHat 8 )
- Hadoop Latest Version
- JDK ( Java Development Kit )
- More than one OS ( You can use VM, Cloud )
That's it let's do this
- First Download Hadoop and JDK
- Now Install this with redhat cmd "rpm -i file(hadoop/jdk)"
- Now Configure Hadoop Core and HDFS file to create Cluster;
After same as this create a client and upload some file and use tcpdump cmd to check uploading of file.
Now we'll see files on WebUI, for this type ip:50070 on browser;
Now we check uploading of file in our slave node for this we use tcpdump -i enp0s3 -n -X
Now Check in your slave node that client IP is coming in your pc to upload file;
If you have more than one slave node and if you stopped one of slave node then also hadoop will upload file because of it's replication feature.
To prove this we uploaded a big file to see replication, block is showing in the pic is the replication of file;
You can perform this practical alone just create more than instance, like this I done this same on AWS Instance;?
Thank you
Consultant at Xebia
4 年Awesome