Home | About | Sematext search-lucene.com search-hadoop.com
 Search Hadoop and all its subprojects:

Switch to Threaded View
MapReduce, mail # user - FSDataOutputStream hangs in out.close()


Copy link to this message
-
Re: FSDataOutputStream hangs in out.close()
Pedro Sá da Costa 2013-03-27, 15:53
Hi,

I'm trying to make the same client to talk to different HDFS and JT
instances that are in different sites of Amazon EC2. The error that I got
is:

 java.io.IOException: Got error for OP_READ_BLOCK,
self=/XXX.XXX.XXX.123:44734,

 remote=ip-XXX-XXX-XXX-123.eu-west-1.compute.internal/XXX.XXX.XXX.123:50010,
for file

 ip-XXX-XXX-XXX-123.eu-west-1.compute.internal/XXX.XXX.XXX.123:50010:-4664365259588027316,
for block
   -4664365259588027316_2050

This error means than it wasn't possible to write on a remote host?

On 27 March 2013 12:24, Harsh J <[EMAIL PROTECTED]> wrote:

> You can try to take a jstack stack trace and see what its hung on.
> I've only ever noticed a close() hang when the NN does not accept the
> complete-file call (due to minimum replication not being guaranteed),
> but given your changes (which I haven't an idea about yet) it could be
> something else as well. You're essentially trying to make the same
> client talk to two different FSes I think (aside of the JT RPC).
>
> On Wed, Mar 27, 2013 at 5:50 PM, Pedro Sá da Costa <[EMAIL PROTECTED]>
> wrote:
> > Hi,
> >
> > I'm using the Hadoop 1.0.4 API to try to submit a job in a remote
> > JobTracker. I created modfied the JobClient to submit the same job in
> > different JTs. E.g, the JobClient is in my PC and it try to submit the
> same
> > Job  in 2 JTs at different sites in Amazon EC2. When I'm launching the
> Job,
> > in the setup phase, the JobClient is trying to submit split file info
> into
> > the remote JT.  This is the method of the JobClient that I've the
> problem:
> >
> >
> >   public static void createSplitFiles(Path jobSubmitDir,
> >       Configuration conf, FileSystem   fs,
> >       org.apache.hadoop.mapred.InputSplit[] splits)
> >   throws IOException {
> >     FSDataOutputStream out = createFile(fs,
> >         JobSubmissionFiles.getJobSplitFile(jobSubmitDir), conf);
> >     SplitMetaInfo[] info = writeOldSplits(splits, out, conf);
> >     out.close();
> >
> >
> writeJobSplitMetaInfo(fs,JobSubmissionFiles.getJobSplitMetaFile(jobSubmitDir),
> >         new FsPermission(JobSubmissionFiles.JOB_FILE_PERMISSION),
> > splitVersion,
> >         info);
> >   }
> >
> > 1 - The FSDataOutputStream hangs in the out.close() instruction. Why it
> > hangs? What should I do to solve this?
> >
> >
> > --
> > Best regards,
>
>
>
> --
> Harsh J
>

--
Best regards,