Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
MapReduce >> mail # user >> Yarn Container Limit


Copy link to this message
-
Re: Yarn Container Limit
You can use yarn.nodemanager.resource.memory-mb to set the limit on
each NodeManager.

You should have a good look at
http://hadoop.apache.org/common/docs/r0.23.0/hadoop-yarn/hadoop-yarn-site/ClusterSetup.html
. It has enough information to get you a good distance.

HTH.
+Vinod

On Tue, Jan 10, 2012 at 3:44 AM, raghavendhra rahul
<[EMAIL PROTECTED]> wrote:
> Hi,
>             How to set the maximum number of containers to be executed in
> each node.
>             So that at a time only that much of containers will be running
> in that node..
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB