I'm more than a little concerned that you missed the whole multiple directories--including a remote one--for the fsimage thing. That's probably the #1 thing that most of the big grids do to maintain the NN data. I can only remember one failure where the NFS copy wasn't used to recover a namenode in all the failures I've personally been involved (and that was an especially odd bug, not a NN failure, per se). The only reason to fall back to the 2ndary NN in 0.20 should be is if you've hit a similarly spectacular bug. Point blank: anyone who runs the NN without it writing to a remote copy doesn't know what they are doing.
Also, until AvatarNode comes of age (which, from what I understand, FB has only been doing for very long themselves), there is no such thing as HA NN. We all have high hopes that it works out, but it likely isn't anywhere near ready for primetime yet.
On Feb 14, 2011, at 2:52 PM, Mark Kerzner wrote:
> I completely agree, and I am using yours and the group's posting to define
> the direction and approaches, but I am also trying every solution - and I am
> beginning to do just that, the AvatarNode now.
> Thank you,
> On Mon, Feb 14, 2011 at 4:43 PM, M. C. Srivas <[EMAIL PROTECTED]> wrote:
>> I understand you are writing a book "Hadoop in Practice". If so, its
>> important that what's recommended in the book should be verified in
>> practice. (I mean, beyond simply posting in this newsgroup - for instance,
>> the recommendations on NN fail-over should be tried out first before
>> about how to do it). Otherwise you won't know your recommendations really
>> work or not.
>> On Mon, Feb 14, 2011 at 12:31 PM, Mark Kerzner <[EMAIL PROTECTED]
>>> Thank you, M. C. Srivas, that was enormously useful. I understand it now,
>>> but just to be complete, I have re-formulated my points according to your
>>> - In 0.20 the Secondary NameNode performs snapshotting. Its data can be
>>> used to recreate the HDFS if the Primary NameNode fails. The procedure
>>> manual and may take hours, and there is also data loss since the last
>>> - In 0.21 there is a Backup Node (HADOOP-4539), which aims to help with
>>> HA and act as a cold spare. The data loss is less than with Secondary
>>> but it is still manual and potentially error-prone, and it takes hours;
>>> - There is an AvatarNode patch available for 0.20, and Facebook runs
>>> cluster that way, but the patch submitted to Apache requires testing
>>> developers adopting it must do some custom configurations and also
>>> care in their work.
>>> As a conclusion, when building an HA HDFS cluster, one needs to follow
>>> practices outlined by Tom
>>> and may still need to resort to specialized NSF filers for running the
>>> On Mon, Feb 14, 2011 at 11:50 AM, M. C. Srivas <[EMAIL PROTECTED]>
>>>> The summary is quite inaccurate.
>>>> On Mon, Feb 14, 2011 at 8:48 AM, Mark Kerzner <[EMAIL PROTECTED]>
>>>>> is it accurate to say that
>>>>> - In 0.20 the Secondary NameNode acts as a cold spare; it can be
>>>>> recreate the HDFS if the Primary NameNode fails, but with the delay
>>>>> minutes if not hours, and there is also some data loss;
>>>> The Secondary NN is not a spare. It is used to augment the work of the
>>>> Primary, by offloading some of its work to another machine. The work
>>>> offloaded is "log rollup" or "checkpointing". This has been a source of
>>>> constant confusion (some named it incorrectly as a "secondary" and now
>>>> are stuck with it).
>>>> The Secondary NN certainly cannot take over for the Primary. It is not