Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
Hive >> mail # user >> Error while reading from task log url


Copy link to this message
-
Re: Error while reading from task log url
After setting this in Hive-

hive> SET hive.exec.show.job.failure.debug.info=false;

I can see the logs on my console itself? Or I need to go somewhere to see
the actual logs and what is causing the problem?
*Raihan Jamal*

On Fri, Jul 20, 2012 at 12:28 PM, [EMAIL PROTECTED] <
[EMAIL PROTECTED]> wrote:

> First of all, this exception is not what is causing your job to fail. When
> a job fails Hive attempts to automatically retrieve the task logs from the
> JobTracker's TaskLogServlet. This indicates something wrong with your
> hadoop setup, JobTracker down, maybe?
>
> You can suppress this exception by doing:
>
> hive> SET hive.exec.show.job.failure.debug.info=false;
>
> Look into your task logs to see why your job actually failed.
>
> On Fri, Jul 20, 2012 at 2:12 PM, Raihan Jamal <[EMAIL PROTECTED]>wrote:
>
>> Whenever I run the below query-
>> *
>> *
>> *SELECT buyer_id, item_id, ranknew(buyer_id, item_id), created_time*
>> *FROM (*
>> *    SELECT buyer_id, item_id, created_time*
>> *    FROM testingtable1*
>> *    DISTRIBUTE BY buyer_id, item_id*
>> *    SORT BY buyer_id, item_id, created_time desc*
>> *) a*
>> *WHERE ranknew(buyer_id, item_id) %  2 == 0;*
>>
>>
>> I always get the below error, I have no clue what does this error means?
>> Is there any problem with my query or something wrong with the system?
>>
>> *Total MapReduce jobs = 1*
>> *Launching Job 1 out of 1*
>> *Number of reduce tasks not specified. Estimated from input data size: 1*
>> *In order to change the average load for a reducer (in bytes):*
>> *  set hive.exec.reducers.bytes.per.reducer=<number>*
>> *In order to limit the maximum number of reducers:*
>> *  set hive.exec.reducers.max=<number>*
>> *In order to set a constant number of reducers:*
>> *  set mapred.reduce.tasks=<number>*
>> *Starting Job = job_201207172005_14407, Tracking URL >> http://ares-jt.vip.ebay.com:50030/jobdetails.jsp?jobid=job_201207172005_14407
>> *
>> *Kill Command = /home/hadoop/latest/bin/../bin/hadoop job
>>  -Dmapred.job.tracker=ares-jt:8021 -kill job_201207172005_14407*
>> *2012-07-21 02:07:15,917 Stage-1 map = 0%,  reduce = 0%*
>> *2012-07-21 02:07:27,211 Stage-1 map = 100%,  reduce = 0%*
>> *2012-07-21 02:07:38,700 Stage-1 map = 100%,  reduce = 33%*
>> *2012-07-21 02:07:48,517 Stage-1 map = 100%,  reduce = 0%*
>> *2012-07-21 02:08:49,566 Stage-1 map = 100%,  reduce = 0%*
>> *2012-07-21 02:09:08,640 Stage-1 map = 100%,  reduce = 100%*
>> *Ended Job = job_201207172005_14407 with errors*
>> *java.lang.RuntimeException: Error while reading from task log url*
>> *        at
>> org.apache.hadoop.hive.ql.exec.errors.TaskLogProcessor.getErrors(TaskLogProcessor.java:130)
>> *
>> *        at
>> org.apache.hadoop.hive.ql.exec.ExecDriver.showJobFailDebugInfo(ExecDriver.java:931)
>> *
>> *        at
>> org.apache.hadoop.hive.ql.exec.ExecDriver.execute(ExecDriver.java:716)*
>> *        at
>> org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:107)*
>> *        at
>> org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:55)
>> *
>> *        at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:621)*
>> *        at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:495)*
>> *        at org.apache.hadoop.hive.ql.Driver.run(Driver.java:374)*
>> *        at
>> org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:138)*
>> *        at
>> org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:197)*
>> *        at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:302)
>> *
>> *        at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)*
>> *        at
>> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:39)
>> *
>> *        at
>> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:25)
>> *
>> *        at java.lang.reflect.Method.invoke(Method.java:597)*
>> *        at org.apache.hadoop.util.RunJar.main(RunJar.java:156)*
>> *Caused by: java.io.IOException: Server returned HTTP response code: 400