Yes it is required -- in psuedodistributed node the jobtracker is not
necessarily aware that the task trackers / data nodes are on the same
machine, and will thus attempt to ssh into them when starting the
respective deamons etc (i.e. start-all.sh)
On Thu, May 16, 2013 at 11:21 AM, kishore alajangi <
[EMAIL PROTECTED]> wrote:
> When you start the hadoop procecess, each process will ask the password to
> start, to overcome this we will configure SSH if you use single node or
> multiple nodes for each process, if you can enter the password for each
> process Its not a mandatory even if you use multiple systems.
> On Thu, May 16, 2013 at 8:24 PM, Raj Hadoop <[EMAIL PROTECTED]> wrote:
>> I have a dedicated user on Linux server for hadoop. I am installing it in
>> psedo distributed mode on this box. I want to test my programs on this
>> machine. But i see that in installation steps - they were mentioned that
>> SSH needs to be configured. If it is single node, I dont require it
>> ...right? Please advise.
>> I was looking at this site
>> It menionted like this -
>> Hadoop requires SSH access to manage its nodes, i.e. remote machines plus
>> your local machine if you want to use Hadoop on it (which is what we want
>> to do in this short tutorial). For our single-node setup of Hadoop, we
>> therefore need to configure SSH access to localhost for the hduser user
>> we created in the previous section.