Ah, yes. Toms book is a good start, and Eric Sammers book Hadoop Operations too :)
BR, AL > On 26 Mar 2015, at 11:50, Mich Talebzadeh <m...@peridale.co.uk> wrote: > > Many thanks AL. I believe you meant “Hadoop the definitive guide” J > > Mich Talebzadeh > > http://talebzadehmich.wordpress.com <http://talebzadehmich.wordpress.com/> > > Publications due shortly: > Creating in-memory Data Grid for Trading Systems with Oracle TimesTen and > Coherence Cache > > NOTE: The information in this email is proprietary and confidential. This > message is for the designated recipient only, if you are not the intended > recipient, you should destroy it immediately. Any information in this message > shall not be understood as given or endorsed by Peridale Ltd, its > subsidiaries or their employees, unless expressly so stated. It is the > responsibility of the recipient to ensure that this email is virus free, > therefore neither Peridale Ltd, its subsidiaries nor their employees accept > any responsibility. > > From: Alexander Alten-Lorenz [mailto:wget.n...@gmail.com] > Sent: 26 March 2015 10:30 > To: user@hadoop.apache.org > Subject: Re: Total memory available to NameNode > > Hi Mich, > > the book Hadoop Operations may a good start: > https://books.google.de/books?id=drbI_aro20oC&pg=PA308&lpg=PA308&dq=hadoop+memory+namenode&source=bl&ots=t_yltgk_i7&sig=_6LXkcSjfuwwqfz_kDGDi9ytgqU&hl=en&sa=X&ei=Nt8TVfn9AcjLPZyXgKAC&ved=0CFYQ6AEwBg#v=onepage&q=hadoop%20memory%20namenode&f=false > > <https://books.google.de/books?id=drbI_aro20oC&pg=PA308&lpg=PA308&dq=hadoop+memory+namenode&source=bl&ots=t_yltgk_i7&sig=_6LXkcSjfuwwqfz_kDGDi9ytgqU&hl=en&sa=X&ei=Nt8TVfn9AcjLPZyXgKAC&ved=0CFYQ6AEwBg#v=onepage&q=hadoop > memory namenode&f=false> > > BR, > AL > > >> On 26 Mar 2015, at 11:16, Mich Talebzadeh <m...@peridale.co.uk >> <mailto:m...@peridale.co.uk>> wrote: >> >> Is there any parameter that sets the total memory that NameNode can use? >> >> Thanks >> >> Mich Talebzadeh >> >> http://talebzadehmich.wordpress.com <http://talebzadehmich.wordpress.com/> >> >> Publications due shortly: >> Creating in-memory Data Grid for Trading Systems with Oracle TimesTen and >> Coherence Cache >> >> NOTE: The information in this email is proprietary and confidential. This >> message is for the designated recipient only, if you are not the intended >> recipient, you should destroy it immediately. Any information in this >> message shall not be understood as given or endorsed by Peridale Ltd, its >> subsidiaries or their employees, unless expressly so stated. It is the >> responsibility of the recipient to ensure that this email is virus free, >> therefore neither Peridale Ltd, its subsidiaries nor their employees accept >> any responsibility. >> >> From: Mirko Kämpf [mailto:mirko.kae...@gmail.com >> <mailto:mirko.kae...@gmail.com>] >> Sent: 25 March 2015 16:08 >> To: user@hadoop.apache.org <mailto:user@hadoop.apache.org>; >> m...@peridale.co.uk <mailto:m...@peridale.co.uk> >> Subject: Re: can block size for namenode be different from wdatanode block >> size? >> >> Correct, let's say you run the NameNode with just 1GB of RAM. >> This would be a very strong limitation for the cluster. For each file we >> need about 200 bytes and for each block as well. Now we can estimate the >> max. capacity depending on HDFS-Blocksize and average File size. >> >> Cheers, >> Mirko >> >> 2015-03-25 15:34 GMT+00:00 Mich Talebzadeh <m...@peridale.co.uk >> <mailto:m...@peridale.co.uk>>: >> Hi Mirko, >> >> Thanks for feedback. >> >> Since i have worked with in memory databases, this metadata caching sounds >> more like an IMDB that caches data at start up from disk resident storage. >> >> IMDBs tend to get issues when the cache cannot hold all data. Is this the >> case the case with metada as well? >> >> Regards, >> >> Mich >> Let your email find you with BlackBerry from Vodafone >> From: Mirko Kämpf <mirko.kae...@gmail.com <mailto:mirko.kae...@gmail.com>> >> Date: Wed, 25 Mar 2015 15:20:03 +0000 >> To: user@hadoop.apache.org >> <mailto:user@hadoop.apache.org><user@hadoop.apache.org >> <mailto:user@hadoop.apache.org>> >> ReplyTo: user@hadoop.apache.org <mailto:user@hadoop.apache.org> >> Subject: Re: can block size for namenode be different from datanode block >> size? >> >> Hi Mich, >> >> please see the comments in your text. >> >> >> >> 2015-03-25 15:11 GMT+00:00 Dr Mich Talebzadeh <m...@peridale.co.uk >> <mailto:m...@peridale.co.uk>>: >> >> Hi, >> >> The block size for HDFS is currently set to 128MB by defauilt. This is >> configurable. >> Correct, an HDFS client can overwrite the cfg-property and define a >> different block size for HDFS blocks. >>> >>> My point is that I assume this parameter in hadoop-core.xml sets the >>> block size for both namenode and datanode. >> Correct, the block-size is a "HDFS wide setting" but in general the >> HDFS-client makes the blocks. >> >>> However, the storage and >>> random access for metadata in nsamenode is different and suits smaller >>> block sizes. >> HDFS blocksize has no impact here. NameNode metadata is held in memory. For >> reliability it is dumped to local discs of the server. >> >>> >>> For example in Linux the OS block size is 4k which means one HTFS blopck >>> size of 128MB can hold 32K OS blocks. For metadata this may not be >>> useful and smaller block size will be suitable and hence my question. >> Remember, metadata is in memory. The fsimage-file, which contains the >> metadata >> is loaded on startup of the NameNode. >> >> Please be not confused by the two types of block-sizes. >> >> Hope this helps a bit. >> Cheers, >> Mirko >> >>> >>> Thanks, >>> >>> Mich