Re: [ANNOUNCE] Hadoop version 1.2.1 (stable) released

2013-08-05 Thread Matt Foley
On Mon, Aug 5, 2013 at 9:43 AM, Chris K Wensel ch...@wensel.net wrote: any particular reason the 1.1.2 releases were pulled from the mirrors (so quickly)? On Aug 4, 2013, at 2:08 PM, Matt Foley ma...@apache.org wrote: I'm happy to announce that Hadoop version 1.2.1 has passed its release

Re: [ANNOUNCE] Hadoop version 1.2.1 (stable) released

2013-08-05 Thread Matt Foley
'stable' link that itself is updated to the actual stable release. we do this now with all of the cascading artifacts. it prevents automated things from breaking immediately. http://files.cascading.org/sdk/2.1/latest.txt ckw On Aug 5, 2013, at 11:20 AM, Matt Foley ma...@apache.org wrote

[ANNOUNCE] Hadoop version 1.2.1 (stable) released

2013-08-04 Thread Matt Foley
I'm happy to announce that Hadoop version 1.2.1 has passed its release vote and is now available. It has 18 bug fixes and patches over the previous 1.2.0 release; please see Hadoop 1.2.1 Release Noteshttp://hadoop.apache.org/docs/r1.2.1/releasenotes.htmlfor details. This release of Hadoop-1.2 is

Re: [ANNOUNCE] Hadoop version 1.2.1 (stable) released

2013-08-04 Thread Matt Foley
which will include Windows native compatibility. My apologies, this was incorrect. Windows has only been integrated to trunk and branch-2.1. Thanks, --Matt On Sun, Aug 4, 2013 at 2:08 PM, Matt Foley ma...@apache.org wrote: I'm happy to announce that Hadoop version 1.2.1 has passed its

Re: Multiple cores vs multiple nodes

2012-07-02 Thread Matt Foley
This is actually a very complex question. Without trying to answer completely, the high points, as I see it, are: a) [Most important] Different kinds of nodes require different Hadoop configurations. In particular, the number of simultaneous tasks per node should presumably be set higher for a

Re: Why is hadoop build I generated from a release branch different from release build?

2012-03-08 Thread Matt Foley
Hi Pawan, The complete way releases are built (for v0.20/v1.0) is documented at http://wiki.apache.org/hadoop/HowToRelease#Building However, that does a bunch of stuff you don't need, like generate the documentation and do a ton of cross-checks. The full set of ant build targets are defined

[ANNOUNCEMENT] Hadoop 0.20.205.0 release

2011-10-18 Thread Matt Foley
On Friday 14 Oct, the Hadoop community voted ten to zero (including four PMC members voting in favor) to accept the release of Hadoop 0.20.205.0. The biggest feature of this release is that it merges the append/hsync/hflush features of branch-0.20-append, and security features of

Re: Hbase with Hadoop

2011-10-13 Thread Matt Foley
Hi Jignesh, the option is --config (with a double dash) not -config (with a single dash). Please let me know if that works. --Matt On Thu, Oct 13, 2011 at 8:30 AM, jigneshmpatel jigneshmpa...@gmail.comwrote: There is no command like -config see below Jignesh-MacBookPro:hadoop-hbase

Re: Hbase with Hadoop

2011-10-12 Thread Matt Foley
Hi Jignesh, Not clear what's going on with your ZK, but as a starting point, the hsync/flush feature in 205 was implemented with an on-off switch. Make sure you've turned it on by setting *dfs.support.append *to true in the hdfs-site.xml config file. Also, are you installing Hadoop with

Re: Hbase with Hadoop

2011-10-11 Thread Matt Foley
Hi Jignesh, 0.20.204.0 does not have hflush/sync support, but 0.20.205.0 does. Without HDFS hsync, HBase will still work, but is subject to data loss if the datanode is restarted. In 205, this deficiency is fixed. 0.20.205.0-rc2 is up for vote in common-dev@. Please try it out with HBase :-)

Re: Version Mismatch

2011-08-18 Thread Matt Foley
Hi Alan, It seems your XXX application incorporates a DFSClient, which implies it is compiled in the presence of certain Hadoop jar files. If it grabs those jar files and incorporates them in the XXX installable package (tarball, rpm, whatever), then it's easy to get this kind of mis-match.

Re: Can i safely set dfs.blockreport.intervalMsec to very large value (1 year or more?)

2011-07-08 Thread Matt Foley
Hi Moon, The periodic block report is constructed entirely from info in memory, so there is no complete scan of the filesystem for this purpose. The periodic block report defaults to only sending once per hour from each datanode, and each DN calculates a random start time for the hourly cycle