I am creating a local HDFS dev environment (actually hasoop + mesos + zk + kafka) to facilitate Spark and facilitate local integrated testing. All other components work fine, but I have problems with HDFS. When the Data Node tries to connect to the node name, I get a DisallowedDataNodeException :
org.apache.hadoop.hdfs.server.protocol.DisallowedDatanodeException: Datanode denied communication with namenode
Most issues related to the same problem boil down to resolving Node data under the Node name, either static through etc/hosts files, or using dns. Static resolution is not an option with docker, since I donβt know the data nodes when creating a container named Node. I would like to avoid creating and maintaining an additional DNS service. Ideally, I would like to connect everything using the --link function of the docker.
Is there a way to configure HDFS to use only IP addresses?
I found this property and set it to false, but this did not help:
dfs.namenode.datanode.registration.ip-hostname-check (default: true)
Is there a way for a local HDFS cluster with multiple nodes to work only using IP addresses and without using DNS?
maasg source share