Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
MapReduce >> mail # user >> Only log.index



Ah, I should've guessed that. You seem to have JVM reuse enabled. Even if JVMs are reused, all the tasks write to the same files as they share the JVM. They only have different index files. The same thing happens for what we call the TaskCleanup tasks which are launched for failing/killed tasks.

Thanks,
+Vinod

On Jul 23, 2013, at 10:55 PM, Ajay Srivastava wrote:

> Hi Vinod,
>
> Thanks. It seems that something else is going on -
>
> Here is the content of log.index -
>
> ajay-srivastava:userlogs ajay.srivastava$ cat job_201307222115_0188/attempt_201307222115_0188_r_000000_0/log.index
> LOG_DIR:/opt/hadoop/bin/../logs/userlogs/job_201307222115_0188/attempt_201307222115_0188_r_000008_0
> stdout:0 0
> stderr:156 0
> syslog:995 166247
>
> Looks like that the log.index is pointing to another attempt directory.
> Is it doing some kind of optimization ? What is purpose of log.index ?
>
>
> Regards,
> Ajay Srivastava
>
>
> On 24-Jul-2013, at 11:09 AM, Vinod Kumar Vavilapalli wrote:
>
>>
>> It could either mean that all those task-attempts are crashing before the process itself is getting spawned (check TT logs) or those logs are getting deleted after the fact. Suspect the earlier.
>>
>> Thanks,
>> +Vinod
>>
>> On Jul 23, 2013, at 9:33 AM, Ajay Srivastava wrote:
>>
>>> Hi,
>>>
>>> I see that most of the tasks have only log.index created in /opt/hadoop/logs/userlogs/jobId/task_attempt directory.
>>> When does this happen ?
>>> Is there a config setting for this OR this is a bug ?
>>>
>>>
>>> Regards,
>>> Ajay Srivastava
>>
>