Re: NN Memory Jumps every 1 1/2 hours

2012-12-27 Thread Suresh Srinivas
is repeated for every file. This in the later release has been optimized. But these optimizations affect only the fsimage and not the memory consumed on the namenode. > > Will I be saving 400,000,000 bytes of memory if I do? > On Thu, Dec 27, 2012 at 5:41 PM, Suresh Srinivas >wrote: >

Re: NN Memory Jumps every 1 1/2 hours

2012-12-27 Thread Suresh Srinivas
ode data has to fit in memory" but what we > really mean is "the name node data must fit in memory 3x" > > On Thu, Dec 27, 2012 at 5:08 PM, Suresh Srinivas >wrote: > > > You did free up lot of old generation with reducing young generation, > > right? The extra

Re: NN Memory Jumps every 1 1/2 hours

2012-12-27 Thread Suresh Srinivas
enough you die a > slow death. > > On Sun, Dec 23, 2012 at 9:40 PM, Suresh Srinivas >wrote: > > > Do not have access to my computer. Based on reading the previous email, I > > do not see any thing suspicious on the list of objects in the histo live > > dump. >

Re: NN Memory Jumps every 1 1/2 hours

2012-12-23 Thread Suresh Srinivas
Dec 22, 2012 at 10:23 PM, Suresh Srinivas > wrote: > >> -XX:NewSize=1G -XX:MaxNewSize=1G

Re: NN Memory Jumps every 1 1/2 hours

2012-12-22 Thread Suresh Srinivas
This looks to me is because of larger default young generation size in newer java releases - see http://docs.oracle.com/javase/6/docs/technotes/guides/vm/cms-6.html#heap_size. I can see looking at your GC logs, around 6G space being used for young generation (though I do not see logs related to min

Re: NN Memory Jumps every 1 1/2 hours

2012-12-22 Thread Suresh Srinivas
Please take a histo live dump when the memory is full. Note that this causes full gc. http://docs.oracle.com/javase/6/docs/technotes/tools/share/jmap.html What are the number of blocks you have on the system. Send the JVM options you are using. From earlier java versions which used 1/8 of total

Re: can "HADOOP-6546: BloomMapFile can return false negatives" get backported to branch-1?

2012-05-08 Thread Suresh Srinivas
This change in merged into branch-1 and will be available in release 1.1. On Mon, May 7, 2012 at 6:40 PM, Jim Donofrio wrote: > Can someone backport HADOOP-6546: BloomMapFile can return false negatives > to branch-1 for the next 1+ release? > > Without this fix BloomMapFile is somewhat useless b

Re: can "HADOOP-6546: BloomMapFile can return false negatives" get backported to branch-1?

2012-05-07 Thread Suresh Srinivas
I have marked it for 1.1. I will follow up on promoting the path. Regards, Suresh On May 7, 2012, at 6:40 PM, Jim Donofrio wrote: > Can someone backport HADOOP-6546: BloomMapFile can return false negatives to > branch-1 for the next 1+ release? > > Without this fix BloomMapFile is somewhat us

Re: Best practice to migrate HDFS from 0.20.205 to CDH3u3

2012-05-03 Thread Suresh Srinivas
This probably is a more relevant question in CDH mailing lists. That said, what Edward is suggesting seems reasonable. Reduce replication factor, decommission some of the nodes and create a new cluster with those nodes and do distcp. Could you share with us the reasons you want to migrate from Apa

Re: hadoop permission guideline

2012-03-22 Thread Suresh Srinivas
Can you please take this discussion CDH mailing list? On Mar 22, 2012, at 7:51 AM, Michael Wang wrote: > I have installed Cloudera hadoop (CDH). I used its Cloudera Manager to > install all needed packages. When it was installed, the root is used. I > found the installation created some users

Re: Issue when starting services on CDH3

2012-03-15 Thread Suresh Srinivas
Guys, can you please take this up in CDH related mailing lists. On Thu, Mar 15, 2012 at 10:01 AM, Manu S wrote: > Because for large clusters we have to run namenode in a single node, > datanode in another nodes > So we can start namenode and jobtracker in master node and datanode n > tasktracker

Re: What is the NEW api?

2012-03-11 Thread Suresh Srinivas
> there are many people talking about the NEW API This might be related to releases 0.21 or later, where append and related functionality is re-implemented. 1.0 comes from 0.20.205 and has same API as 0.20-append. Sent from phone On Mar 11, 2012, at 6:27 PM, WangRamon wrote: > > > > > Hi

Re: Backupnode in 1.0.0?

2012-02-23 Thread Suresh Srinivas
On Thu, Feb 23, 2012 at 12:41 AM, Jeremy Hansen wrote: > Thanks. Could you clarify what BackupNode does? > > -jeremy > > Namenode currently keeps the entire file system namespace in memory. It logs the write operations (create, delete file etc.) into a journal file called editlog. This journal n

Re: Backupnode in 1.0.0?

2012-02-22 Thread Suresh Srinivas
at 2:43 PM, Joey Echeverria wrote: > >> > >>> Check out the Apache Bigtop project. I believe they have 0.22 RPMs. > >>> > >>> Out of curiosity, why are you interested in BackupNode? > >>> > >>> -Joey > >>> > >>>

Re: Backupnode in 1.0.0?

2012-02-22 Thread Suresh Srinivas
BackupNode is major functionality with change in required in RPC protocols, configuration etc. Hence it will not be available in bug fix release 1.0.1. It is also unlikely to be not available on minor releases in 1.x release streams. Regards, Suresh On Wed, Feb 22, 2012 at 11:40 AM, Jeremy Hanse

Re: datanode failing to start

2012-01-09 Thread Suresh Srinivas
Can you please send your notes on what info is out of date or better still create a jira so that it can be addressed. On Fri, Jan 6, 2012 at 3:11 PM, Dave Kelsey wrote: > gave up and installed version 1. > it installed correctly and worked, thought the instructions for setup and > the location o

Re: HDFS Backup nodes

2011-12-13 Thread Suresh Srinivas
Srivas, As you may know already, NFS is just being used in the first prototype for HA. Two options for editlog store are: 1. Using BookKeeper. Work has already completed on trunk towards this. This will replace need for NFS to store the editlogs and is highly available. This solution will also b

Re: Difference between DFS Used and Non-DFS Used

2011-07-08 Thread Suresh Srinivas
non DFS storage is not required, it is provided as information only to shown how the storage is being used. The available storage on the disks is used for both DFS and non DFS (mapreduce shuffle output and any other files that could be on the disks). See if you have unnecessary files or shuffle o

Re: CDH and Hadoop

2011-03-24 Thread suresh srinivas
On Thu, Mar 24, 2011 at 7:04 PM, Rita wrote: > Oh! Thats for the heads up on that... > > I guess I will go with the cloudera source then > > > On Thu, Mar 24, 2011 at 8:41 PM, David Rosenstrauch >wrote: > > > They do, but IIRC, they recently announced that they're going to be > > discontinuing i

Re: Data Nodes do not start

2011-02-09 Thread suresh srinivas
On Tue, Feb 8, 2011 at 11:05 PM, rahul patodi wrote: > I think you should copy the namespaceID of your master which is in > name/current/VERSION file to all the slaves > This is a sure recipe for disaster. The VERSION file is a file system meta data file not to be messed around with. At worst, t