One way: This may be possible to do if you inflate the
"dfs.datanode.du.reserved" property on the specific DataNode to a very
large bytes value (> maximum volume size). This way your NN will still
consider the DN as a valid one that carries readable blocks, but when
writing files this DN will never be selected due to its
On Mon, May 21, 2012 at 12:37 AM, Kyungyong Lee <[EMAIL PROTECTED]> wrote:
> I would like to ask if I can do the following. Assuming that I have a
> datanode, i.e., dn1, which already contains some useful blocks. Here,
> I do not want to save new data blocks to the datanode, but I still
> want to use the blocks that already exist in the datanode (dn1).
> I considered to use exclude file (dfs.hosts.exclude). However, if I
> add "dn1" to the exclude file list, I cannot use blocks that are
> already contained in dn1. If it is right, can you please give me some
> guidances to do what I'm thinking using HDFS?