Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Plain View
HBase, mail # user - Question about MapReduce


+
Jean-Marc Spaggiari 2012-10-27, 20:30
+
Jean-Marc Spaggiari 2012-10-29, 15:11
+
Shrijeet Paliwal 2012-10-29, 17:03
+
Jean-Marc Spaggiari 2012-11-02, 18:56
+
Shrijeet Paliwal 2012-11-02, 19:06
+
Jean-Marc Spaggiari 2012-11-02, 19:31
+
Jean-Marc Spaggiari 2012-11-02, 19:47
+
Shrijeet Paliwal 2012-11-02, 19:51
+
Jean-Marc Spaggiari 2012-11-02, 20:01
Copy link to this message
-
Re: Question about MapReduce
Jean-Marc Spaggiari 2012-11-14, 02:41
One more question about MapReduce.

One of my servers is slower than the others. I don't have any time
constraint for the job to finish.

But I'm getting this message:

"Task attempt_201211122318_0014_m_000021_0 failed to report status for
601 seconds. Killing!"

Where can I chance this timeout to something like 1800 seconds? Is it
on the mapred-site.xml file? If so, which property should I insert?

Thanks,

JM

2012/11/2, Jean-Marc Spaggiari <[EMAIL PROTECTED]>:
> That was my initial plan too, but I was wondering if there was any
> other best practice about the delete.  So I will go that way.
>
> Thanks,
>
> JM
>
> 2012/11/2, Shrijeet Paliwal <[EMAIL PROTECTED]>:
>> Not sure what exactly is happening in your job. But in one of the delete
>> jobs I wrote I was creating an instance of HTable in setup method of my
>> mapper
>>
>> delTab = new HTable(conf, conf.get(TABLE_NAME));
>>
>> And performing delete in map() call using delTab. So no, you do not have
>> access to table directly *usually*.
>>
>>
>> -Shrijeet
>>
>>
>> On Fri, Nov 2, 2012 at 12:47 PM, Jean-Marc Spaggiari <
>> [EMAIL PROTECTED]> wrote:
>>
>>> Sorry, one last question.
>>>
>>> On the map method, I have access to the row using the values
>>> parameter. Now, based on the value content, I might want to delete it.
>>> Do I have access to the table directly from one of the parameters? Or
>>> should I call the delete using an HTableInterface from my pool?
>>>
>>> Thanks,
>>>
>>> JM
>>>
>>> 2012/11/2, Jean-Marc Spaggiari <[EMAIL PROTECTED]>:
>>> > Yep, you perfectly got my question.
>>> >
>>> > I just tried and it's working perfectly!
>>> >
>>> > Thanks a lot! I now have a lot to play with.
>>> >
>>> > JM
>>> >
>>> > 2012/11/2, Shrijeet Paliwal <[EMAIL PROTECTED]>:
>>> >> JM,
>>> >>
>>> >> I personally would chose to put it neither hadoop libs nor hbase
>>> >> libs.
>>> >> Have
>>> >> them go to your application's own install directory.
>>> >>
>>> >> Then you could sent the variable HADOOP_CLASSPATH to have your jar
>>> >> (also
>>> >> include hbase jars, hbase dependencies and dependencies your program
>>> >> needs)
>>> >> And to execute fire 'hadoop jar' command.
>>> >>
>>> >> An example[1]:
>>> >>
>>> >> Set classpath:
>>> >> export HADOOP_CLASSPATH=`hbase
>>> classpath`:mycool.jar:mycooldependency.jar
>>> >>
>>> >> Fire following to launch your job:
>>> >> hadoop jar mycool.jar hbase.experiments.MyCoolProgram
>>> >> -Dmapred.running.map.limit=50
>>> >> -Dmapred.map.tasks.speculative.execution=false aCommandLineArg
>>> >>
>>> >>
>>> >> Did I get your question right?
>>> >>
>>> >> [1] In the example I gave `hbase classpath` gets you set with all
>>> >> hbase
>>> >> jars.
>>> >>
>>> >>
>>> >>
>>> >> On Fri, Nov 2, 2012 at 11:56 AM, Jean-Marc Spaggiari <
>>> >> [EMAIL PROTECTED]> wrote:
>>> >>
>>> >>> Hi Shrijeet,
>>> >>>
>>> >>> Helped a lot! Thanks!
>>> >>>
>>> >>> Now, the only think I need is to know where's the best place to put
>>> >>> my
>>> >>> JAR on the server. Should I put it on the hadoop lib directory? Or
>>> >>> somewhere on the HBase structure?
>>> >>>
>>> >>> Thanks,
>>> >>>
>>> >>> JM
>>> >>>
>>> >>> 2012/10/29, Shrijeet Paliwal <[EMAIL PROTECTED]>:
>>> >>> > In line.
>>> >>> >
>>> >>> > On Mon, Oct 29, 2012 at 8:11 AM, Jean-Marc Spaggiari <
>>> >>> > [EMAIL PROTECTED]> wrote:
>>> >>> >
>>> >>> >> I'm replying to myself ;)
>>> >>> >>
>>> >>> >> I found "cleanup" and "setup" methods from the TableMapper table.
>>> >>> >> So
>>> >>> >> I
>>> >>> >> think those are the methods I was looking for. I will init the
>>> >>> >> HTablePool there. Please let me know if I'm wrong.
>>> >>> >>
>>> >>> >> Now, I still have few other questions.
>>> >>> >>
>>> >>> >> 1) context.getCurrentValue() can throw a InterrruptedException,
>>> >>> >> but
>>> >>> >> when can this occur? Is there a timeout on the Mapper side? Of
>>> >>> >> it's
>>> >>> >> if
>>> >>> >> the region is going down while the job is running?
>>> >>> >>
>>> >>> >