I'm not sure if I understand your question correctly, but I think this is exactly what the standard workflow in a Hadoop application looks like:
Job job1 = new Job(...);
// setup job, set Mapper and Reducer, etc.
job1.waitForCompletion(...); // at this point, the cluster will run job 1 and wait for its completion
// think about the results from job 1, plan job 2 accordingly
Job job2 = new Job(...);
// setup another job
job2.waitForCompletion(...); // at this point, the cluster will run job 2 and wait for its completion
So I think "waitForCompletion" is what you were asking about, right?
Von: Hamid Oliaei [mailto:[EMAIL PROTECTED]]
Gesendet: Dienstag, 26. Juni 2012 11:00
An: [EMAIL PROTECTED]
Betreff: Understanding job completion in other nodes
I want to run a job on all of nodes and if one job was completed, the node must wait until the jobs on the other nodes finish.
For that, every node must signal to all nodes and when every node receives the signal from every one, next job must be run.
How can I handle that in Hadoop?
Is there any solution to understanding job completion in other nodes?
P.S: I am using Hadoop 0.20.2.