Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
HBase >> mail # user >> Exceptions in Hadoop and Hbase log files


Copy link to this message
-
Exceptions in Hadoop and Hbase log files
Hi,
I am running Hbase in pseudo distributed mode.( Hbase 0.94.7 and Hadoop
1.1.2).
I am getting certain exceptions in Hadoop's namenode and data node files
which are :-

Namenode :-

2013-10-18 10:33:37,218 INFO org.apache.hadoop.hdfs.StateChange: BLOCK*
NameSystem.heartbeatCheck: lost heartbeat from 192.168.20.30:50010
2013-10-18 10:33:37,242 INFO org.apache.hadoop.net.NetworkTopology:
Removing a node: /default-rack/192.168.20.30:50010
2013-10-18 10:35:27,606 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Number of
transactions: 64 Total time for transactions(ms): 1Number
of transactions batched in Syncs: 0 Number of syncs: 43 SyncTimes(ms): 86
2013-10-18 10:35:27,614 ERROR
org.apache.hadoop.security.UserGroupInformation: PriviledgedActionException
as:hadoop cause:java.io.IOException: File /h
base/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
could only be replicated to 0 nodes, instead of 1
2013-10-18 10:35:27,895 INFO org.apache.hadoop.ipc.Server: IPC Server
handler 9 on 9000, call
addBlock(/hbase/event_data/433b61f2a4ebff8f2e4b89890508a
3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e,
DFSClient_hb_rs_hbase.rummycircle.com,60020,1382012725057, null) from
192.168.20.30:44990: error: java.io.I
OException: File
/hbase/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
could only be replicated to 0 nodes, instead
 of 1
java.io.IOException: File
/hbase/event_data/433b61f2a4ebff8f2e4b89890508a3b7/.tmp/99797a61a8f7471cb6df8f7b95f18e9e
could only be replicated to 0 nodes
, instead of 1
    at
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.getAdditionalBlock(FSNamesystem.java:1639)
    at
org.apache.hadoop.hdfs.server.namenode.NameNode.addBlock(NameNode.java:736)
    at sun.reflect.GeneratedMethodAccessor10.invoke(Unknown Source)
    at
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
    at java.lang.reflect.Method.invoke(Method.java:597)
    at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:578)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1393)
    at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1389)
    at java.security.AccessController.doPrivileged(Native Method)
    at javax.security.auth.Subject.doAs(Subject.java:396)
    at
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1149)
    at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1387)
Data node :-

2013-10-18 06:13:14,499 WARN
org.apache.hadoop.hdfs.server.datanode.DataNode: DatanodeRegistration(
192.168.20.30:50010, storageID=DS-1816106352-192.16
8.20.30-50010-1369314076237, infoPort=50075, ipcPort=50020):Got exception
while serving blk_-3215981820534544354_52215 to /192.168.20.30:
java.net.SocketTimeoutException: 480000 millis timeout while waiting for
channel to be ready for write. ch :
java.nio.channels.SocketChannel[connected
 local=/192.168.20.30:50010 remote=/192.168.20.30:36188]
    at
org.apache.hadoop.net.SocketIOWithTimeout.waitForIO(SocketIOWithTimeout.java:246)
    at
org.apache.hadoop.net.SocketOutputStream.waitForWritable(SocketOutputStream.java:159)
    at
org.apache.hadoop.net.SocketOutputStream.transferToFully(SocketOutputStream.java:198)
    at
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendChunks(BlockSender.java:392)
    at
org.apache.hadoop.hdfs.server.datanode.BlockSender.sendBlock(BlockSender.java:490)
    at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:202)
    at
org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:104)
    at java.lang.Thread.run(Thread.java:662)

--
Thanks and Regards,
Vimal Jain
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB