请解释什么是dfs.include文件目的以及如何定义它.
我已经向Hadoop集群添加了一个新节点,但它没有被namenode标识.在其中一篇文章中,我发现dfs.include可以解决此问题.
先感谢您,
Vladi 最佳答案 仅在dfs.include和mapred.include中包含节点名称是不够的.必须在namenode / jobtracker上更新从属文件.必须在新节点上启动tasktracker和datanode,并且必须在NameNode和JobTracker上运行refreshNodes命令,以使它们知道新节点.
以下是关于如何做到这一点的instructions.
根据‘Hadoop : The Definitive Guide’
The file (or files) specified by the dfs.hosts and mapred.hosts properties is different from the slaves file. The former is used by the namenode and jobtracker to determine which worker nodes may connect. The slaves file is used by the Hadoop control scripts to perform cluster-wide operations, such as cluster restarts. It is never used by the Hadoop
daemons.