Pal Konyves 2013-04-20, 17:52
Ted Yu 2013-04-20, 20:07
Only one family, my data is very simple key-value, although I want to make
sequential scan, so making a hash of the key is not an option.
On Sat, Apr 20, 2013 at 10:07 PM, Ted Yu <[EMAIL PROTECTED]> wrote:
> How many column families do you have ?
> For #3, per-splitting table at the row keys corresponding to peaks makes
> On Apr 20, 2013, at 10:52 AM, Pal Konyves <[EMAIL PROTECTED]> wrote:
> > Hi,
> > I am just reading about region splitting. By default - as I understand -
> > Hbase handles splitting the regions. I just don't know how to imagine on
> > which key it splits the regions.
> > 1) For example when I write MD5 hash of rowkeys, they are most probably
> > evenly distributed from
> > 000000... to FFFFF... right? When Hbase starts with one region, all the
> > writes goes into that region, and when the HFile get's too big, it just
> > gets for example the median value of the stored keys, and split the
> > by this?
> > 2) I want to bulk load tons of data with the HBase java client API put
> > operations. I want it to perform well. My keys are numeric sequential
> > values (which I know from this post, I cannot load into Hbase
> > because the Hbase tables are going to be sad
> > )
> > So I thought I would pre-split the table into regions, and load the data
> > randomized. This way I will get good distribution among region servers in
> > terms of network IO from the beginning. Is that a good idea?
> > 3) If my rowkeys are not evenly distributed in the keyspace, but they
> > some peaks or bursts. e.g. 000-999, but most of the keys gather around
> > and 060 values, is it a good idea to have the pre region splits at those
> > peaks?
> > Thanks in advance,
> > Pal
Ted Yu 2013-04-20, 20:54
Pal Konyves 2013-04-20, 21:24
Ted Yu 2013-04-21, 01:34
Pal Konyves 2013-04-21, 11:21