Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
MapReduce, mail # user - Re: Incompartible cluserIDS


Copy link to this message
-
Re: Incompartible cluserIDS
Mohammad Tariq 2013-04-29, 22:09
make it 755.

Warm Regards,
Tariq
https://mtariq.jux.com/
cloudfront.blogspot.com
On Tue, Apr 30, 2013 at 3:30 AM, Kevin Burton <[EMAIL PROTECTED]>wrote:

> Thank you the HDFS system seems to be up. Now I am having a problem with
> getting the JobTracker and TaskTracker up. According to the logs on the
> JobTracker mapred doesn't have  write permission to /. I am not clear on
> what the permissions should be.
>
> Anyway, thank you.
>
> On Apr 29, 2013, at 4:30 PM, Mohammad Tariq <[EMAIL PROTECTED]> wrote:
>
> Hello Kevin,
>
>           Have you reformatted the NN(unsuccessfully)?Was your NN serving
> some other cluster earlier or your DNs were part of some other
> cluster?Datanodes bind themselves to namenode through namespaceID and in
> your case the IDs of DNs and NN seem to be different. As a workaround you
> could do this :
>
> 1- Stop all the daemons.
> 2- Go to the directory which you have specified as the value of
> "dfs.name.dir" property in your hdfs-site.xml file.
> 3- You'll find a directory called "current" inside this directory where a
> file named "VERSION" will be present. Open this file and copy the value of
> "namespaceID" form here.
> 4- Now go to the directory which you have specified as the value of
> "dfs.data.dir" property in your hdfs-site.xml file.
> 5- Move inside the "current" directory and open the "VERSION" file here as
> well. Now replace the value of "namespaceID" present here with the one you
> had copied earlier.
> 6- Restart all the daemons.
>
> Note : If you have not created dfs.name.dir and dfs.data.dir separately,
> you could find all this inside your temp directory.
>
> HTH
>
> Warm Regards,
> Tariq
> https://mtariq.jux.com/
> cloudfront.blogspot.com
>
>
> On Tue, Apr 30, 2013 at 2:45 AM, <[EMAIL PROTECTED]> wrote:
>
>> I am trying to start up a cluster and in the datanode log on the NameNode
>> server I get the error:
>>
>> 2013-04-29 15:50:20,988 INFO
>> org.apache.hadoop.hdfs.server.common.Storage: Lock on
>> /data/hadoop/dfs/data/in_use.lock acquired by nodename 1406@devUbuntu05
>> 2013-04-29 15:50:20,990 FATAL
>> org.apache.hadoop.hdfs.server.datanode.DataNode: Initialization failed for
>> block pool Block pool BP-1306349046-172.16.26.68-1367256199559 (storage id
>> DS-403514403-172.16.26.68-50010-1366406077018) service to devUbuntu05/
>> 172.16.26.68:9000
>> java.io.IOException: *Incompatible clusterIDs* in /data/hadoop/dfs/data:
>> namenode clusterID = CID-23b9f9c7-2c25-411f-8bd2-4d5c9d7c25a1; datanode
>> clusterID = CID-e3f6b811-c1b4-4778-a31e-14dea8b2cca8
>>         at
>> org.apache.hadoop.hdfs.server.datanode.DataStorage.doTransition(DataStorage.java:391)
>>         at
>> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:191)
>>         at
>> org.apache.hadoop.hdfs.server.datanode.DataStorage.recoverTransitionRead(DataStorage.java:219)
>>
>> How do I get around this error? What does the error mean?
>>
>> Thank you.
>>
>> Kevin
>>
>
>