Mark Olimpiati 2013-09-11, 17:09
I assume you are talking about MapReduce. And 1.x release or 2.x?
In either of the releases, this cannot be done directly.
In 1.x, the framework doesn't expose a feature like this as it is a shared service, and if enough jobs flock to a node, it will lead to utilization and failure handling issues.
In Hadoop 2 YARN, the platform does expose this functionality. But MapReduce framework doesn't yet expose this functionality to the end users.
What exactly is your use case? Why are some nodes of higher priority than others?
+Vinod Kumar Vavilapalli
On Sep 11, 2013, at 10:09 AM, Mark Olimpiati wrote:
> Thanks for replying Rev, but the link is talking about reducers which seems to be like a similar case but what if I assigned priorities to the data partitions (eg. partition B=1, partition C=2, partition A=3,...) such that first map task is assigned partition B to run first. Then second map is given partition C, .. etc. This is instead of assigning based on partition size. Is that possible?
> On Mon, Sep 9, 2013 at 11:17 AM, Ravi Prakash <[EMAIL PROTECTED]> wrote:
> From: Mark Olimpiati <[EMAIL PROTECTED]>
> To: [EMAIL PROTECTED]
> Sent: Friday, September 6, 2013 1:47 PM
> Subject: assign tasks to specific nodes
> Hi guys,
> I'm wondering if there is a way for me to assign tasks to specific machines or at least assign priorities to the tasks to be executed in that order. Any suggestions?
NOTICE: This message is intended for the use of the individual or entity to
which it is addressed and may contain information that is confidential,
privileged and exempt from disclosure under applicable law. If the reader
of this message is not the intended recipient, you are hereby notified that
any printing, copying, dissemination, distribution, disclosure or
forwarding of this communication is strictly prohibited. If you have
received this communication in error, please contact the sender immediately
and delete it from your system. Thank You.