Hadoop shutdown scripts failing

2014-09-29 Thread Susheel Kumar Gadalay
How to redirect the storing of the following files from /tmp to some other location. hadoop-os user-namenode.pid hadoop-os user-datanode.pid yarn-os user-resourcemanager.pid yarn-os user-nodemanager.pid In /tmp, these files are cleared by OS sometime back and I am unable to shutdown by standard

Re: No space when running a hadoop job

2014-09-29 Thread Aitor Cedres
Hi Susheel, Adding a new directory to “dfs.datanode.data.dir” will not balance your disks straightforward. Eventually, by HDFS activity (deleting/invalidating some block, writing new ones), the disks will become balanced. If you want to balance them right after adding the new disk and changing

Re: Hadoop shutdown scripts failing

2014-09-29 Thread Aitor Cedres
Hi Susheel, You have to set in your hadoop-env.sh and yarn-env.sh the variabkes: - HADOOP_PID_DIR - YARN_PID_DIR To point to some other directory (most common is /var/run/hadoop-hdfs hadoop-yarn) Hope it helps, Aitor On 29 September 2014 07:50, Susheel Kumar Gadalay skgada...@gmail.com wrote:

How to overwrite container-log4j.properties file

2014-09-29 Thread Hanish Bansal
Hi All, I want to use custom logging for map-reduce application so i want to configure my log4j.properties file in hadoop. I want that logging of all containers should be happen according to my log4.properties file. For this i updated different properties in mapred-site.xml configuration file.

Re: Hadoop shutdown scripts failing

2014-09-29 Thread Susheel Kumar Gadalay
Thanks Aitor. There is a shell variable HADOOP_PID_DIR refererred in hadoop-env.sh but not in yarn-env.sh. On 9/29/14, Aitor Cedres aced...@pivotal.io wrote: Hi Susheel, You have to set in your hadoop-env.sh and yarn-env.sh the variabkes: - HADOOP_PID_DIR - YARN_PID_DIR To point to

Re: No space when running a hadoop job

2014-09-29 Thread Susheel Kumar Gadalay
Thank Aitor. That is what is my observation too. I added a new disk location and manually moved some files. But if 2 locations are given at the beginning itself for dfs.datanode.data.dir, will hadoop balance the disks usage, if not perfect because file sizes may differ. On 9/29/14, Aitor

Extremely amount of memory and DB connections by MR Job

2014-09-29 Thread Blanca Hernandez
Hi, I am using a hadoop map reduce job + mongoDb. It goes against a data base 252Gb big. During the job the amount of conexions is over 8000 and we gave already 9Gb RAM. The job is still crashing because of a OutOfMemory with only a 8% of the mapping done. Are this numbers normal? Or did we

Re: No space when running a hadoop job

2014-09-29 Thread Aitor Cedres
I think they way it works when HDFS has a list in dfs.datanode.data.dir, it's basically a round robin between disks. And yes, it may not be perfect balanced cause of different file sizes. On 29 September 2014 13:15, Susheel Kumar Gadalay skgada...@gmail.com wrote: Thank Aitor. That is what

Re: Hadoop shutdown scripts failing

2014-09-29 Thread Aitor Cedres
Check the file $HADOOP_HOME/bin/yarn-daemon.sh; there is a reference to YARN_PID_DIR. If it's not set. it will default to /tmp. On 29 September 2014 13:11, Susheel Kumar Gadalay skgada...@gmail.com wrote: Thanks Aitor. There is a shell variable HADOOP_PID_DIR refererred in hadoop-env.sh but

RE: Extremely amount of memory and DB connections by MR Job

2014-09-29 Thread java8964
I don't have any experience with MongoDB, but just gave my 2 cents here. Your code is not efficient, as using the += on String, and you could have reused the Text object in your mapper, as it is a mutable class, to be reused and avoid creating it again and again like new Text() in the mapper. My

AW: Extremely amount of memory and DB connections by MR Job

2014-09-29 Thread Blanca Hernandez
Thanks for your answer. To your questions: 1. When you claim 96G ram, I am not sure what do you mean? It is not 96 Gb RAM, it is 9 Gb that our test server has available (is it too small?). 2. Your code is not efficient, as using the += on String I need (or at least I don´t have

RE: Re: Regarding HDFS and YARN support for S3

2014-09-29 Thread Naganarasimha G R (Naga)
Hi Takenori, Thanks for replying but still seem not getting some concepts I understand that we need to give fs.AbstractFileSystem.s3.impl if we want to submit job using ./yarn jar with S3 HCFS configured. But what i don't understand is why 2 interfaces (may be i am novice in HDFS and hence not

Re: Re: Regarding HDFS and YARN support for S3

2014-09-29 Thread Takenori Sato
Hi Naga, But what i don't understand is why 2 interfaces (may be i am novice in HDFS and hence not able to completely correlate with jira's which you gave). A client program is encouraged to use FileContext API instead of FileSystem API. Here's why

RE: AW: Extremely amount of memory and DB connections by MR Job

2014-09-29 Thread java8964
Here are my suggestions originally aims to improve the efficient: 1) In your case, you could use StringBuilder, which has the append method, should be more efficient to concatenate your string data in this case.2) What I mean to reuse the Text object is as following: public class mapper

Using Yarn in end to end tests

2014-09-29 Thread Alex Newman
I am currently developing tests that use a mini yarn cluster. Because it is running on circle-ci I need to use the absolute minimum amount of memory. I'm currently setting conf.setFloat(yarn. nodemanager.vmem-pmem-ratio, 8.0f); conf.setBoolean(mapreduce.map.speculative, false);

Re: Hadoop shutdown scripts failing

2014-09-29 Thread Susheel Kumar Gadalay
Thanks On 9/29/14, Aitor Cedres aced...@pivotal.io wrote: Check the file $HADOOP_HOME/bin/yarn-daemon.sh; there is a reference to YARN_PID_DIR. If it's not set. it will default to /tmp. On 29 September 2014 13:11, Susheel Kumar Gadalay skgada...@gmail.com wrote: Thanks Aitor. There is a

binding namenode and job tracker to 0.0.0.0

2014-09-29 Thread Bharath Kumar
Hi , I have 2 different networks in my setup Job tracker and name node are running in private network, eclipse client is running on public network I see the JIRA which is very relevant is there a workaround ? https://issues.apache.org/jira/browse/HADOOP-1202 -- Warm Regards,

Re: No space when running a hadoop job

2014-09-29 Thread Abdul Navaz
Dear All, I am not doing load balancing here. I am just copying a file and it is throwing me an error no space left on the device. hduser@dn1:~$ df -h Filesystem Size Used Avail Use% Mounted on /dev/xvda2 5.9G 5.1G

YARN application failing to localize needed jars (Xpost: hbase users)

2014-09-29 Thread iain wright
Hi folks, I'm having trouble using HBASE copyTable to seed an existing tables data to a replication peer. Surely its an oversight in configuration on our part, but I've scoured the web and doc's for a couple days now. We have been able to run these jobs with success (perhaps they don't require

Kerberosed Hadoop HDFS

2014-09-29 Thread Xiaohua Chen
Hi Experts: I write the following java program to access Kerberosed Hadoop File system: --- import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileStatus;

Re: From java application : how to access kerberosed hadoop HDFS ?

2014-09-29 Thread Xiaohua Chen
Resolved: from client machine I need to update the ./etc/hadoop/core-site.xml with site specific info e.g. property namehadoop.security.authentication/name valuekerberos/value /property also hdfs-site.xml needs: property namedfs.namenode.kerberos.principal/name

how to access oozie web console for kerberosed hadoop

2014-09-29 Thread Xiaohua Chen
Hi Experts: I can go to oozie console http://oozieserver:11000/oozie/ when our hadoop cluster has no kerberos setup. But after Keberos setup, I try the same oozie console from IE, i got error: type Status report message

Re: Hadoop UI - Unable to connect to the application master from the Hadoop UI.

2014-09-29 Thread S.L
The host name is fully qualified , meaning there is nothing more that I can add , it just seems the ports might be messed up , but I don't know which ones On Mon, Sep 29, 2014 at 12:44 AM, Susheel Kumar Gadalay skgada...@gmail.com wrote: I also faced some issue like this. It shows the URL in

Re: Failed to active namenode when config HA

2014-09-29 Thread ??????
Hi, Matt Thank you very much for your response! There were some mistakes in my description as i wrote this mail in a hurry. I put those properties is in hdfs-site.xml not core-site.xml. There are four name nodes because i also using HDFS federation, so there are two nameservices in porperty

RE: Failed to active namenode when config HA

2014-09-29 Thread Brahma Reddy Battula
You need to start the ZKFC process which will monitor and manage the state of namenode. Automatic failover adds two new components to an HDFS deployment: a ZooKeeper quorum, and the ZKFailoverController process (abbreviated as ZKFC). Apache ZooKeeper is a highly available service for