Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
HBase >> mail # user >> HBase issues since upgrade from 0.92.4 to 0.94.6


Copy link to this message
-
Re: HBase issues since upgrade from 0.92.4 to 0.94.6
Thank you for your responses. With respect to the version of Java I found
that Cloudera recommend<http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH4/latest/CDH4-Requirements-and-Supported-Versions/cdhrsv_topic_3.html>1.7.x
for CDH4.3.
On Fri, Jul 12, 2013 at 1:32 PM, Jean-Marc Spaggiari <
[EMAIL PROTECTED]> wrote:

> Might want to run memtest also, just to be sure there is no memory issue.
> It should not since it was working fine with 0.92.4, but costs nothing...
>
> the last version of Java 6 is 45... Might also worst to give it a try if
> you are running with 1.6.
>
> 2013/7/12 Asaf Mesika <[EMAIL PROTECTED]>
>
> > You need to see the jvm crash in .out log file and see if maybe its the
> .so
> > native Hadoop code that making the problem. In our case we
> > Downgraded from jvm 1.6.0-37 to 33 and it solved the issue.
> >
> >
> > On Friday, July 12, 2013, David Koch wrote:
> >
> > > Hello,
> > >
> > > NOTE: I posted the same message in the the Cloudera group.
> > >
> > > Since upgrading from CDH 4.0.1 (HBase 0.92.4) to 4.3.0 (HBase 0.94.6)
> we
> > > systematically experience problems with region servers crashing
> silently
> > > under workloads which used to pass without problems. More specifically,
> > we
> > > run about 30 Mapper jobs in parallel which read from HDFS and insert in
> > > HBase.
> > >
> > > region server log
> > > NOTE: no trace of crash, but server is down and shows up as such in
> > > Cloudera Manager.
> > >
> > > 2013-07-12 10:22:12,050 WARN
> > > org.apache.hadoop.hbase.regionserver.wal.HLogSplitter: File
> > >
> > >
> >
> hdfs://XXXXXXX:8020/hbase/.logs/XXXXXXX,60020,1373616547696-splitting/XXXXXXX%2C60020%2C1373616547696.1373617004286
> > > might be still open, length is 0
> > > 2013-07-12 10:22:12,051 INFO org.apache.hadoop.hbase.util.FSHDFSUtils:
> > > Recovering file
> > >
> > >
> >
> hdfs://XXXXXXX:8020/hbase/.logs/XXXXXXX,60020,1373616547696-splitting/XXXXXXX
> > > t%2C60020%2C1373616547696.1373617004286
> > > 2013-07-12 10:22:13,064 INFO org.apache.hadoop.hbase.util.FSHDFSUtils:
> > > Finished lease recover attempt for
> > >
> > >
> >
> hdfs://XXXXXXX:8020/hbase/.logs/XXXXXXX,60020,1373616547696-splitting/XXXXXXX%2C60020%2C1373616547696.1373617004286
> > > 2013-07-12 10:22:14,819 INFO org.apache.hadoop.io.compress.CodecPool:
> Got
> > > brand-new compressor [.deflate]
> > > 2013-07-12 10:22:14,824 INFO org.apache.hadoop.io.compress.CodecPool:
> Got
> > > brand-new compressor [.deflate]
> > > ...
> > > 2013-07-12 10:22:14,850 INFO org.apache.hadoop.io.compress.CodecPool:
> Got
> > > brand-new compressor [.deflate]
> > > 2013-07-12 10:22:15,530 INFO org.apache.hadoop.io.compress.CodecPool:
> Got
> > > brand-new compressor [.deflate]
> > > < -- last log entry, region server is down here -- >
> > >
> > >
> > > datanode log, same machine
> > >
> > > 2013-07-12 10:22:04,811 ERROR
> > > org.apache.hadoop.hdfs.server.datanode.DataNode:
> > XXXXXXX:50010:DataXceiver
> > > error processing WRITE_BLOCK operation  src: /YYY.YY.YYY.YY:36024 dest:
> > > /XXX.XX.XXX.XX:50010
> > > java.io.IOException: Premature EOF from inputStream
> > > at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:194)
> > > at
> > >
> > >
> >
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213)
> > > at
> > >
> > >
> >
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
> > > at
> > >
> > >
> >
> org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
> > > at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:414)
> > > at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:635)
> > > at
> > >
> > >
> >
> org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:564)
> > > at
> > >
> > >
> >
> org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:103)
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB