Actually coprocessor can be used to implement MR-like function, while not
using Hadoop framework.
Research Staff Member
IBM T. J. Watson Research Center
Yorktown Heights, NY 10598
[EMAIL PROTECTED]; 914-784-6752
From: Dalia Sobhy <[EMAIL PROTECTED]>
To: "[EMAIL PROTECTED]" <[EMAIL PROTECTED]>,
Date: 11/24/2012 01:33 PM
Subject: RE: Hbase MapReduce
It 's weird that hbase aggregate functions don't use MapReduce, this means
that the performance will be very poor.
Is it a must to use coprocessors?
Is there a much easier way to improve the functions' performance ?
> CC: [EMAIL PROTECTED]
> From: [EMAIL PROTECTED]
> Subject: Re: Hbase MapReduce
> Date: Sat, 24 Nov 2012 12:05:45 -0600
> To: [EMAIL PROTECTED]
> Do you think it would be a good idea to temper the use of CoProcessors?
> This kind of reminds me of when people first started using stored
> Sent from a remote device. Please excuse any typos...
> Mike Segel
> On Nov 24, 2012, at 11:46 AM, tom <[EMAIL PROTECTED]> wrote:
> > Hi, but you do not need to us M/R. You could also use coprocessors.
> > See this site:
> > https://blogs.apache.org/hbase/entry/coprocessor_introduction
> > -> in the section "Endpoints"
> > An aggregation coprocessor ships with hbase that should match your
> > You just need to load it and eventually you can access it from HTable:
> > HTable.coprocessorExec(..) <
> > Regards
> > tom
> > Am 24.11.2012 18:32, schrieb Marcos Ortiz:
> >> Regards, Dalia.
> >> You have to use MapReduce for that.
> >> In the HBase in Practice´s book, there are lot of great examples for
> >> On 11/24/2012 12:15 PM, Dalia Sobhy wrote:
> >>> Dear all,
> >>> I wanted to ask a question..
> >>> Do Hbase Aggregate Functions such as rowcount, getMax, get Average
use MapReduce to execute those functions?
> >>> Thanks :D