RE: OOM Error Map output copy.

2011-12-08 Thread Devaraj K
Hi Niranjan, Every thing looks ok as per the info you have given. Can you check in the job.xml file whether these child opts reflecting or any thing else is overwriting this config. 3. mapred.child.java.opts --> -Xms512M -Xmx1536M -XX:+UseSerialGC and also can you tell me which

RE: HDFS Backup nodes

2011-12-08 Thread Jorn Argelo - Ephorus
Hi Koji, This was on CHD3U1. For the record I had the dfs.name.dir.restore which Harsh mentioned enabled as well. Jorn -Oorspronkelijk bericht- Van: Koji Noguchi [mailto:knogu...@yahoo-inc.com] Verzonden: woensdag 7 december 2011 17:59 Aan: common-user@hadoop.apache.org Onderwerp: Re: H

Re: how to read hadoop source code efficiently

2011-12-08 Thread Praveen Sripati
Harsh, Are you mentioning about the 0.23 or pre-0.23 release? The OP was about 0.23 release. Were you able to start some daemons in Eclipse and submit a job? Some time back I posted a query [1] about setting 0.23 in Eclipse and the response was that no one might have tried it. I gave it up after

Routing and region deletes

2011-12-08 Thread Per Steffensen
Hi The system we are going to work on will receive 50mio+ new datarecords every day. We need to keep a history of 2 years of data (thats 35+ billion datarecords in the storage all in all), and that basically means that we also need to delete 50mio+ datarecords every day, or e.g. 1,5 billion e

Re: Routing and region deletes

2011-12-08 Thread Michel Segel
Per Seffensen, I would urge you to step away from the keyboard and rethink your design. It sounds like you want to replicate a date partition model similar to what you would do if you were attempting this with HBase. HBase is not a relational database and you have a different way of doing things

Re: Routing and region deletes

2011-12-08 Thread Per Steffensen
Thanks for your reply! Michel Segel skrev: Per Seffensen, I would urge you to step away from the keyboard and rethink your design. Will do :-) But would actually still like to receive answers for my questions - just pretend that my ideas are not so stupid and let me know if it can be done

Re: OOM Error Map output copy.

2011-12-08 Thread Niranjan Balasubramanian
Devaraj These are indeed the actual settings I copied over from the job.xml. ~ Niranjan. On Dec 8, 2011, at 12:10 AM, Devaraj K wrote: > Hi Niranjan, > > Every thing looks ok as per the info you have given. Can you check > in the job.xml file whether these child opts reflecting or any th

Re: OOM Error Map output copy.

2011-12-08 Thread Niranjan Balasubramanian
I am using version 0.20.203. Thanks ~ Niranjan. On Dec 8, 2011, at 9:26 AM, Niranjan Balasubramanian wrote: > Devaraj > > These are indeed the actual settings I copied over from the job.xml. > > ~ Niranjan. > On Dec 8, 2011, at 12:10 AM, Devaraj K wrote: > >> Hi Niranjan, >> >> Every th

Cloudera Free

2011-12-08 Thread Bai Shen
Does anyone know of a good tutorial for Cloudera Free? I found installation instructions, but there doesn't seem to be in formation on how to run jobs, etc, once you have it set up. Thanks.

Question about accessing another HDFS

2011-12-08 Thread Frank Astier
Hi - We have two namenodes set up at our company, say: hdfs://A.mycompany.com hdfs://B.mycompany.com >From the command line, I can do: Hadoop fs –ls hdfs://A.mycompany.com//some-dir And Hadoop fs –ls hdfs://B.mycompany.com//some-other-dir I’m now trying to do the same from a Java program tha

Re: Question about accessing another HDFS

2011-12-08 Thread Tom Melendez
I'm hoping there is a better answer, but I'm thinking you could load another configuration file (with B.company in it) using Configuration, grab a FileSystem obj with that and then go forward. Seems like some unnecessary overhead though. Thanks, Tom On Thu, Dec 8, 2011 at 2:42 PM, Frank Astier

Re: Question about accessing another HDFS

2011-12-08 Thread Jay Vyas
Can you show your code here ? What URL protocol are you using ? On Thu, Dec 8, 2011 at 5:47 PM, Tom Melendez wrote: > I'm hoping there is a better answer, but I'm thinking you could load > another configuration file (with B.company in it) using Configuration, > grab a FileSystem obj with that a

Re: Question about accessing another HDFS

2011-12-08 Thread Frank Astier
Can you show your code here ? What URL protocol are you using ? I’m guess I’m being very naïve (and relatively new to HDFS). I can’t show too much code, but basically, I’d like to do: Path myPath = new Path(“hdfs://A.mycompany.com//some-dir”); Where Path is a hadoop fs path. I think I can take

Re: Cloudera Free

2011-12-08 Thread Joey Echeverria
Hi Bai, I'm moving this over to scm-us...@cloudera.org as that's a more appropriate list. (common-user bcced). I assume by "Cloudera Free" you mean Coudera Manager Free Edition? You should be able to run a job in the same way that do on any other Hadoop cluster. The only caveat is that you first

Re: Question about accessing another HDFS

2011-12-08 Thread JAX
I was confused about this for a while also I dont have all the details but I think my question on s.o. might help you. I was playing with different protocols... Trying to find a way to programatically access all data in Hfds. http://stackoverflow.com/questions/7844458/how-can-i-access-hadoop

Regarding Parrallel Iron's claim

2011-12-08 Thread JS Jang
Hi, Does anyone know any discussion in Apache Hadoop regarding the claim by Parrallel Iron with their patent against use of HDFS? Thanks in advance. Regards, JS

Re: Regarding Parrallel Iron's claim

2011-12-08 Thread Jean-Daniel Cryans
Isn't that old news? http://www.dbms2.com/2011/06/10/patent-nonsense-parallel-ironhdfs-edition/ Googling around, doesn't seem anything happened after that. J-D On Thu, Dec 8, 2011 at 6:52 PM, JS Jang wrote: > Hi, > > Does anyone know any discussion in Apache Hadoop regarding the claim by > Par

how to integrate snappy into hadoop 0.20.205.0(apache release)

2011-12-08 Thread Jinyan Xu
Hi all, Can anyone tell me how to integrate snappy into hadoop 0.20.205.0(apache release)? Not cloudera version. Thanks! The information and any attached documents contained in this message may be confidential and/or legally privileged. The message is intended s

Re: Regarding Parrallel Iron's claim

2011-12-08 Thread JS Jang
I appreciate your help, J-D. Yes, I wondered whether there was any update since or previous discussion within Apache Hadoop as I am new in this mailing list. On 12/9/11 12:19 PM, Jean-Daniel Cryans wrote: Isn't that old news? http://www.dbms2.com/2011/06/10/patent-nonsense-parallel-ironhdfs-e

Re: Regarding Parrallel Iron's claim

2011-12-08 Thread Jean-Daniel Cryans
You could just look at the archives: http://mail-archives.apache.org/mod_mbox/hadoop-common-user/ It is also indexed by all search engines. J-D On Thu, Dec 8, 2011 at 7:44 PM, JS Jang wrote: > I appreciate your help, J-D. > Yes, I wondered whether there was any update since or previous discussi

Re: Regarding Parrallel Iron's claim

2011-12-08 Thread JS Jang
Got it. Thanks again, J-D. On 12/9/11 12:54 PM, Jean-Daniel Cryans wrote: > You could just look at the archives: > http://mail-archives.apache.org/mod_mbox/hadoop-common-user/ > > It is also indexed by all search engines. > > J-D > > On Thu, Dec 8, 2011 at 7:44 PM, JS Jang wrote: >> I appreciate

Re: Routing and region deletes

2011-12-08 Thread Per Steffensen
Ahhh stupid me. I probably just want to use different tables for different days/months. Believe tables can fairly quickly be deleted on HBase? Regards, Per Steffensen Per Steffensen skrev: Thanks for your reply! Michel Segel skrev: Per Seffensen, I would urge you to step away from the keyb

Re: Not able to post a job in Hadoop 0.23.0

2011-12-08 Thread Arun C Murthy
Moving to mapreduce-user@, bcc common-user@. Can you see any errors in the logs? Typically this happens when you have no NodeManagers. Check the 'nodes' link and then RM logs. Arun On Nov 29, 2011, at 8:36 PM, Nitin Khandelwal wrote: > HI , > > I have successfully setup Hadoop 0.23.0 in a si

Choosing IO intensive and CPU intensive workloads

2011-12-08 Thread ArunKumar
Hi guys ! I want to see the behavior of a single node of Hadoop cluster when IO intensive / CPU intensive workload and mix of both is submitted to the single node alone. These workloads must stress the nodes. I see that TestDFSIO benchmark is good for IO intensive workload. 1> Which benchmarks d

Re: Not able to post a job in Hadoop 0.23.0

2011-12-08 Thread Nitin Khandelwal
Hi Arun, Thanks for your reply. There is one NodeManager running ; Following is the NodeManager UI : Rack Node State Node Address Node HTTP Address Health-status Last health-update Health-report Containers Mem Used Mem Avail /default-rack RUNNING germinait93:50033 germinait93: Healthy 9-Dec

Re: Not able to post a job in Hadoop 0.23.0

2011-12-08 Thread Nitin Khandelwal
CC : mapreduce-user On 9 December 2011 13:14, Nitin Khandelwal wrote: > Hi Arun, > Thanks for your reply. > > > There is one NodeManager running ; Following is the NodeManager UI : > > Rack > Node State > Node Address > Node HTTP Address > Health-status > Last health-update > Health-report > Cont

Re: Choosing IO intensive and CPU intensive workloads

2011-12-08 Thread alo alt
Hi Arun, Micheal has write up a good tutorial about, including stress test and IO. http://www.michael-noll.com/blog/2011/04/09/benchmarking-and-stress-testing-an-hadoop-cluster-with-terasort-testdfsio-nnbench-mrbench/ - Alex On Fri, Dec 9, 2011 at 8:24 AM, ArunKumar wrote: > Hi guys ! > > I wa