Oops, missed the part where you already tried that.
On Mon, Jun 2, 2008 at 3:23 PM, Michael Di Domenico [EMAIL PROTECTED]
wrote:
Depending on your windows version, there is a dos command called subst
which you could use to virtualize a drive letter on your third machine
On Fri, May 30, 2008 at 4:35 AM, Sridhar Raman [EMAIL PROTECTED]
wrote:
Should the installation paths be the same in all the nodes? Most
documentation seems to suggest that it is _*recommended*_ to have the
_*same
*_ paths in all the nodes. But what is the workaround, if, because of
some
reason, one isn't able to have the same path?
That's the problem we are facing right now. After making Hadoop work
perfectly in a 2-node cluster, when we tried to accommodate a 3rd machine,
we realised that this machine doesn't have a E:, which is where the
installation of hadoop is in the other 2 nodes. All our machines are
Windows machines. The possible solutions are:
1) Move the installations in M1 M2 to a drive that is present in M3. We
will keep this as the last option.
2) Map a folder in M3's D: to E:. We used the subst command to do this.
But when we tried to start DFS, it wasn't able to find the hadoop
installation. Just to verify, we tried a ssh to the localhost, and were
unable to find the mapped drive. It's only visible as a folder of D:.
Whereas, in the basic cygwin prompt, we are able to view E:.
3) Partition M3's D drive and create an E. This carries the risk of loss
of
data.
So, what should we do? Is there any way we can specify in the NameNode
the
installation paths of hadoop in each of the remaining nodes? Or is there
some environment variable that can be set, which can make the hadoop
installation path specific to each machine?
Thanks,
Sridhar