On Thursday, March 17, 2011, Marc Sturlese <[EMAIL PROTECTED]> wrote:
> Is there any way to check if a seqfile is corrupted without iterate over all
> its keys/values till it crashes?
> I've seen that I can get an IOException when opening it or an IOException
> reading the X key/value (depending on when it was corrupted).
> Thanks in advance
> View this message in context: http://lucene.472066.n3.nabble.com/check-if-a-sequenceFile-is-corrupted-tp2693230p2693230.html
> Sent from the Hadoop lucene-users mailing list archive at Nabble.com.
No it does not seem possible to know if a file is damaged without
reading it (logically).
Hadoop dfs -text xxxxxx ; echo $?
Should give you non zero but that still involves reading the file.