Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
HBase >> mail # user >> Consequent deletes more than ~256 rows not working


Copy link to this message
-
Re: Consequent deletes more than ~256 rows not working
It would be better paste a piece of your code.
On Tue, Dec 3, 2013 at 7:04 AM, Ted Yu <[EMAIL PROTECTED]> wrote:

> Which HBase release are you using ?
>
> In your while loop, you used the same set of row keys for each attempt ?
>
> Thanks
>
>
> On Sun, Dec 1, 2013 at 11:28 PM, Mrudula Madiraju <
> [EMAIL PROTECTED]
> > wrote:
>
> > Hi,
> >
> > I have a simple hbase table of approx. 1000 rows.
> > If I invoke the htable.delete() on this table in a while loop - it
> doesn't
> > throw any error or exception.
> > But at the end of operation - I see that it has actually deleted only
> > about 256 rows.
> > Repeating the operation deletes another 256 or so. And finally after 3 or
> > 4 runs , all rows finally get deleted.
> >
> > This is true of htable.batch API or even htable.delete API.
> >
> > Have tried changing ulimit/nproc settings, invoking flush, setting
> > autocommit, invoking major Compact
> > Also waited out - That is to see if after 5 minutes of first run delete
> > will complete in background. Nothing works.
> >
> > Searched the mailing list and see that there are threads on delete
> > followed by put not working etc.
> > But this is a different case.
> >
> > Anyone who knows what's happening? Looking forward to any pointers!
> >
> > Regards,
> > Mrudula
>
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB