Actually, I should amend my statement -- SSH is required, but passwordless
ssh (i guess) you can live without if you are willing to enter your
password for each process that gets started.
But Why wouldn't you want to implement passwordless ssh in a pseudo
distributed cluster ? Its very easy to implement on a single node:
cat ~/.ssh/id_rsa.pub /root/.ssh/authorized_keys
On Thu, May 16, 2013 at 11:31 AM, Jay Vyas <[EMAIL PROTECTED]> wrote:
> Yes it is required -- in psuedodistributed node the jobtracker is not
> necessarily aware that the task trackers / data nodes are on the same
> machine, and will thus attempt to ssh into them when starting the
> respective deamons etc (i.e. start-all.sh)
> On Thu, May 16, 2013 at 11:21 AM, kishore alajangi <
> [EMAIL PROTECTED]> wrote:
>> When you start the hadoop procecess, each process will ask the password
>> to start, to overcome this we will configure SSH if you use single node or
>> multiple nodes for each process, if you can enter the password for each
>> process Its not a mandatory even if you use multiple systems.
>> On Thu, May 16, 2013 at 8:24 PM, Raj Hadoop <[EMAIL PROTECTED]> wrote:
>>> I have a dedicated user on Linux server for hadoop. I am installing it
>>> in psedo distributed mode on this box. I want to test my programs on this
>>> machine. But i see that in installation steps - they were mentioned that
>>> SSH needs to be configured. If it is single node, I dont require it
>>> ...right? Please advise.
>>> I was looking at this site
>>> It menionted like this -
>>> Hadoop requires SSH access to manage its nodes, i.e. remote machines
>>> plus your local machine if you want to use Hadoop on it (which is what we
>>> want to do in this short tutorial). For our single-node setup of Hadoop, we
>>> therefore need to configure SSH access to localhost for the hduser user
>>> we created in the previous section.
> Jay Vyas