I've never dealt with a large cluster, though I'd imagine it is managed the same way as small clusters:
-Use hostnames or ips, whichever is more convenient for you -All the slaves need to go into the slave file -You can update software by using bin/hadoop-daemons.sh. Something like: #bin/hadoop-daemons.sh "rsync (mastersrcpath) (localdestpath)" I created a wiki page that currently contains one tip for managing large clusters. Could others add to this wiki page? <http://wiki.apache.org/hadoop/LargeClusterTips> Thanks. Hope this helps! Alex On Thu, Sep 11, 2008 at 5:15 PM, 叶双明 <[EMAIL PROTECTED]> wrote: > Hi, all! > > How to manage a large cluster, eg. more than 2000 nodes. > How to config hostname and ip, use DNS? > How to config slaves, all in slaves file? > How to update software in all nodes. > > Any practice, articles, suggestion is appreciate! > Thanks. > > -- > Sorry for my english!! 明 > Please help me to correct my english expression and error in syntax >