FAQ
Sounds like just a bit more work on understanding ssh will get you there.

What you are looking for is getting that public key into authorized_keys

James

Sent from my mobile. Please excuse the typos.
On 2011-03-04, at 2:58 AM, MANISH SINGLA wrote:

Hii all,
I am trying to setup a 2 node cluster...I have configured all the
files as specified in the tutorial I am refering to...I copied the
public key to the slave's machine...but when I ssh to the slave from
the master, it asks for password everytime...kindly help...
On Fri, Mar 4, 2011 at 11:12 AM, icebergs wrote:
You can check the logs whose tasktracker isn't up.
The path is "HADOOP_HOME/logs/".
The answer may be in it.

2011/3/2 bikash sharma <sharmabiks.07@gmail.com>
Hi Sonal,
Thanks. I guess you are right. ps -ef exposes such processes.

-bikash
On Tue, Mar 1, 2011 at 1:29 PM, Sonal Goyal wrote:

Bikash,

I have sometimes found hanging processes which jps does not report, but a
ps -ef shows them. Maybe you can check this on the errant nodes..

Thanks and Regards,
Sonal
<https://github.com/sonalgoyal/hiho>Hadoop ETL and Data Integration<
https://github.com/sonalgoyal/hiho>
Nube Technologies <http://www.nubetech.co>

<http://in.linkedin.com/in/sonalgoyal>






On Tue, Mar 1, 2011 at 7:37 PM, bikash sharma <sharmabiks.07@gmail.com
wrote:
Hi James,
Sorry for the late response. No, the same problem persists. I
reformatted
HDFS, stopped mapred and hdfs daemons and restarted them (using
start-dfs.sh
and start-mapred.sh from master node). But surprisingly out of 4 nodes
cluster, two nodes have TaskTracker running while other two do not have
TaskTrackers on them (verified using jps). I guess since I have the
Hadoop
installed on shared storage, that might be the issue? Btw, how do I
start
the services independently on each node?

-bikash
On Sun, Feb 27, 2011 at 11:05 PM, James Seigel wrote:

.... Did you get it working? What was the fix?

Sent from my mobile. Please excuse the typos.
On 2011-02-27, at 8:43 PM, Simon wrote:

Hey Bikash,

Maybe you can manually start a tasktracker on the node and see if
there
are
any error messages. Also, don't forget to check your configure files
for
mapreduce and hdfs and make sure datanode can start successfully
first.
After all these steps, you can submit a job on the master node and
see
if
there are any communication between these failed nodes and the
master
node.
Post your error messages here if possible.

HTH.
Simon -

On Sat, Feb 26, 2011 at 10:44 AM, bikash sharma <
sharmabiks.07@gmail.com
wrote:
Thanks James. Well all the config. files and shared keys are on a
shared
storage that is accessed by all the nodes in the cluster.
At times, everything runs fine on initialization, but at other
times,
the
same problem persists, so was bit confused.
Also, checked the TaskTracker logs on those nodes, there does not
seem
to
be
any error.

-bikash

On Sat, Feb 26, 2011 at 10:30 AM, James Seigel <james@tynt.com>
wrote:
Maybe your ssh keys aren’t distributed the same on each machine or
the
machines aren’t configured the same?

J

On 2011-02-26, at 8:25 AM, bikash sharma wrote:

Hi,
I have a 10 nodes Hadoop cluster, where I am running some
benchmarks
for
experiments.
Surprisingly, when I initialize the Hadoop cluster
(hadoop/bin/start-mapred.sh), in many instances, only some nodes
have
TaskTracker process up (seen using jps), while other nodes do not
have
TaskTrackers. Could anyone please explain?

Thanks,
Bikash


--
Regards,
Simon

Search Discussions

Discussion Posts

Previous

Follow ups

Related Discussions

Discussion Navigation
viewthread | post
posts ‹ prev | 10 of 12 | next ›
Discussion Overview
groupcommon-user @
categorieshadoop
postedFeb 26, '11 at 3:26p
activeMar 12, '11 at 4:31p
posts12
users6
websitehadoop.apache.org...
irc#hadoop

People

Translate

site design / logo © 2022 Grokbase