0

If I am using a Hadoop cluster that consists of three nodes with Apache Hadoop version 2.7.1

(active name node nn1,stand by name node nn2 and data node n1)

and we decide to add a new data node to the cluster while it's running dn2 (we call that adding data node on the fly)

What is the best way to achieve that without stopping services on name node

Of course first of all we have to install Hadoop files on this new data node or we can copy them from name node, don't we ?

lmiguelvargasf
  • 63,191
  • 45
  • 217
  • 228
oula alshiekh
  • 843
  • 5
  • 14
  • 40
  • 1
    Possible duplicate of [Is there a way to add nodes to a running Hadoop cluster?](http://stackoverflow.com/questions/13159184/is-there-a-way-to-add-nodes-to-a-running-hadoop-cluster) – jedijs May 08 '17 at 10:40
  • 1
    if you are using Ambari to manage your cluster, this is fairly straightforward (Hosts->Actions) – tbone May 08 '17 at 12:56
  • no iam using apache hadoop 2.7.1 – oula alshiekh May 08 '17 at 13:05
  • thanks jedijs ,yes it is duplicate of this question http://stackoverflow.com/questions/13159184/is-there-a-way-to-add-nodes-to-a-running-hadoop-cluster – oula alshiekh May 18 '17 at 06:02
  • Ambari is a web UI to manage Hadoop clusters. As mentioned, adding a datanode is only a few clicks – OneCricketeer Jun 18 '17 at 07:09

1 Answers1

0

Install and configure the new hadoop node. Make sure that you mention the namenode IP correct in core-site.xml.j2

Then you can start the datanode using the following command

/sbin/hadoop-daemon.sh start datanode

This will add the node to the cluster.

shahin
  • 3,515
  • 1
  • 17
  • 17