Multi Node Configuring a Hadoop Cluster

Pseudo-distributed node cluster implementation

I am using window 7 with CYGWIN and have successfully installed hasoop-1.0.3 . I still run the task tracking, task tracking and namenode services on the port (localhost: 50030, localhost: 50060 and localhost: 50070). I completed the implementation of one node.

Now I want to implement Pseudo-Distributed multiple cluster node. I donโ€™t understand how to separate the master and slave system through the ips network?

+4
source share
2 answers

For your ssh problem, just follow the link of one cluster node:

http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-single-node-cluster/

and yes, you need to specify ip master and slave in the conf file for this you can refer to this URL: http://www.michael-noll.com/tutorials/running-hadoop-on-ubuntu-linux-multi-node- cluster /

Hope this helps.

+1
source

Try creating the virtual machine number that you want to add to your cluster. Make sure that these virtual machines have the same hadoop version. Find out the IP addresses of each virtual machine. you will find files with the name master and slaves in $ HADOOP_HOME / conf that indicate the IP address of the virtual machine on the conf / master file that you want to process as the master, and do the same with conf / slaves with the subordinate IP nodes.

Make sure that these nodes have a connection without a password. Format your namenode and then run start -all.sh.

Thanks,

0
source

All Articles