Home | About | Sematext search-lucene.com search-hadoop.com
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB
 Search Hadoop and all its subprojects:

Switch to Threaded View
MapReduce >> mail # user >> Distributing the code to multiple nodes


Copy link to this message
-
Re: Distributing the code to multiple nodes
German,

This does not seem to be helping. I tried to use the Fairscheduler as my
resource manger but the behavior remains same. I could see the
fairscheduler log getting continuous heart beat from both the other nodes.
But it is still not distributing the work to other nodes. What I did next
was started 3 jobs simultaneously so that may be some part of one of the
job be distributed to other nodes. However still only one node is being
used :(((. What is that is going wrong can some one help?

Sample of fairsheduler log:
2014-01-13 15:13:54,293 HEARTBEAT       l1dev-211
2014-01-13 15:13:54,953 HEARTBEAT       l1-dev06
2014-01-13 15:13:54,988 HEARTBEAT       l1-DEV05
2014-01-13 15:13:55,295 HEARTBEAT       l1dev-211
2014-01-13 15:13:55,956 HEARTBEAT       l1-dev06
2014-01-13 15:13:55,993 HEARTBEAT       l1-DEV05
2014-01-13 15:13:56,297 HEARTBEAT       l1dev-211
2014-01-13 15:13:56,960 HEARTBEAT       l1-dev06
2014-01-13 15:13:56,997 HEARTBEAT       l1-DEV05
2014-01-13 15:13:57,299 HEARTBEAT       l1dev-211
2014-01-13 15:13:57,964 HEARTBEAT       l1-dev06
2014-01-13 15:13:58,001 HEARTBEAT       l1-DEV05

My Data distributed as blocks to other nodes. The host with IP 10.12.11.210
has all the data and this is the one which is serving all the request.

Total number of blocks: 8
1073741866:         10.12.11.211:50010    View Block Info
10.12.11.210:50010    View Block Info
1073741867:         10.12.11.211:50010    View Block Info
10.12.11.210:50010    View Block Info
1073741868:         10.12.11.210:50010    View Block Info
10.12.11.209:50010    View Block Info
1073741869:         10.12.11.210:50010    View Block Info
10.12.11.209:50010    View Block Info
1073741870:         10.12.11.211:50010    View Block Info
10.12.11.210:50010    View Block Info
1073741871:         10.12.11.210:50010    View Block Info
10.12.11.209:50010    View Block Info
1073741872:         10.12.11.211:50010    View Block Info
10.12.11.210:50010    View Block Info
1073741873:         10.12.11.210:50010    View Block Info
10.12.11.209:50010    View Block Info

Someone please advice on how to go about this.

--Ashish
On Fri, Jan 10, 2014 at 12:58 PM, Ashish Jain <[EMAIL PROTECTED]> wrote:

> Thanks for all these suggestions. Somehow I do not have access to the
> servers today and will try the suggestions made on monday and will let you
> know how it goes.
>
> --Ashish
>
>
> On Thu, Jan 9, 2014 at 7:53 PM, German Florez-Larrahondo <
> [EMAIL PROTECTED]> wrote:
>
>> Ashish
>>
>> Could this be related to the scheduler you are using and its settings?.
>>
>>
>>
>> On lab environments when running a single type of job I often use
>> FairScheduler (the YARN default in 2.2.0 is CapacityScheduler) and it does
>> a good job distributing the load.
>>
>>
>>
>> You could give that a try (
>> https://hadoop.apache.org/docs/current/hadoop-yarn/hadoop-yarn-site/FairScheduler.html
>> )
>>
>>
>>
>> I think just changing yarn-site.xml  as follows could demonstrate this
>> theory (note that  how the jobs are scheduled depend on resources such as
>> memory on the nodes and you would need to setup yarn-site.xml accordingly).
>>
>>
>>
>> <property>
>>
>>   <name>yarn.resourcemanager.scheduler.class</name>
>>
>>
>> <value>org.apache.hadoop.yarn.server.resourcemanager.scheduler.fair.FairScheduler</value>
>>
>> </property>
>>
>>
>>
>> Regards
>>
>> ./g
>>
>>
>>
>>
>>
>> *From:* Ashish Jain [mailto:[EMAIL PROTECTED]]
>> *Sent:* Thursday, January 09, 2014 6:46 AM
>> *To:* [EMAIL PROTECTED]
>> *Subject:* Re: Distributing the code to multiple nodes
>>
>>
>>
>> Another point to add here 10.12.11.210 is the host which has everything
>> running including a slave datanode. Data was also distributed this host as
>> well as the jar file. Following are running on 10.12.11.210
>>
>> 7966 DataNode
>> 8480 NodeManager
>> 8353 ResourceManager
>> 8141 SecondaryNameNode
>> 7834 NameNode
>>
>>
>>
>> On Thu, Jan 9, 2014 at 6:12 PM, Ashish Jain <[EMAIL PROTECTED]> wrote:
>>
>> Logs were updated only when I copied the data. After copying the data
NEW: Monitor These Apps!
elasticsearch, apache solr, apache hbase, hadoop, redis, casssandra, amazon cloudwatch, mysql, memcached, apache kafka, apache zookeeper, apache storm, ubuntu, centOS, red hat, debian, puppet labs, java, senseiDB