Re: How to specify the job to run on the specific nodes(machines) in the hadoop yarn cluster?

2014-07-30 Thread Ravi Mutyala
https://issues.apache.org/jira/browse/YARN-796 Not yet released, so this is not yet supported. On Wed, Jul 30, 2014 at 2:34 AM, adu dujinh...@hzduozhun.com wrote: Hi all, RT. I want to run a job on specific two nodes in the cluster? How to configure the yarn? Dose yarn queue help? Thanks

Re: No such file or directory

2014-07-29 Thread Ravi Mutyala
Check if there is /home/hduser (hdfs dfs -ls /home/hduser). I think it does not exist. Hadoop 2 does not create a subdirectory if parent directory does not exist. Also, look for /user which is where hdfs looks for 'home' directories for users. On Tue, Jul 29, 2014 at 3:12 PM, Bhupathi,

Re: hadoop multinode, only master node doing the work

2014-07-13 Thread Ravi Mutyala
Try cleaning up /home/xisisu/mydata/hdfs/datanode on the slave nodes andvstart 'datanode' on that again. That should work. If it doesn't, send the log message. Also see if nodemanagers are running on slaves nodes. If everything is running on master, then its not just HDFS issue, its also yarn

Re: hadoop multinode, only master node doing the work

2014-07-13 Thread Ravi Mutyala
http://localhost:8088/cluster/nodes and see if you have nodemanagers on all nodes (or just the master). If nodemanager on slaves is not running, try starting it. And look in /var/log/hadoop-yarn/yarn/yarn-yarn-nodemanager*.log (or wherever your yarn logs are configured) to see if there are any

Re: hadoop multinode, only master node doing the work

2014-07-13 Thread Ravi Mutyala
, sleepTime=1 SECONDS) I googled and found this: http://hortonworks.com/community/forums/topic/retrying-connect-to-server-0-0-0-00-0-0-08030/ Seems it is the configuration error? Thanks very much! Sisu On Sun, Jul 13, 2014 at 2:33 PM, Ravi Mutyala r...@hortonworks.com wrote: http

Re: Hadoop migration downtime: Hadoop 1.x to 2.x

2014-01-20 Thread Ravi Mutyala
If you have a mirror of your production environment (not in size but in versions of hardware and software), I suggest you run migration steps on this environment to test and measure the down times. Actual down times are dependent on components in use. You could also do some of your backup steps

Re: Removing Dead Nodes from Cluster

2013-08-21 Thread Ravi Mutyala
Check on name node ui for missing blocks. You will lose all the blocks that have replicas only on the lost data nodes. On Aug 19, 2013 4:30 AM, Manickam P manicka...@outlook.com wrote: Hello, I have a hadoop cluster setup with 1 master node and 6 data nodes. I have set the replication factor

Re: hadoop mahout

2013-07-06 Thread Ravi Mutyala
The code samples use hadoop 0.20.2 as i see it in pom.xml. So there is a version mismatch between the client libs and hadoop cluster. See if you can change the dependences to 1.1.2 and see if it builds. On Jul 6, 2013, at 3:22 PM, Manal Helal wrote: Hi, I installed hadoop 1.1.2 on 4

Re: Getting started recommendations

2013-01-11 Thread Ravi Mutyala
On Fri, Jan 11, 2013 at 4:29 AM, John Lilley john.lil...@redpoint.netwrote: Where would we find some “big data” files that people have used for testing purposes? Some of the most commonly used 'Big Data' files for testing are Global Weather Data from NCDC

Re: HDFS disk space requirement

2013-01-10 Thread Ravi Mutyala
If the file is a txt file, you could get a good compression ratio. Changing the replication to 3 and the file will fit. But not sure what your usecase is what you want to achieve by putting this data there. Any transformation on this data and you would need more space to save the transformed data.