Re: How to split DBInputFormat?

2011-01-04 Thread Hari Sreekumar
Thanks Sonal. I'll look into this tool as well. hari On Tue, Jan 4, 2011 at 3:57 PM, Sonal Goyal wrote: > Hi Hari, > > I dont think DataDrivenDBInputFormat is available in 0.20.x, its only > available in 0.21 versions. You can check hihoApache0.20 branch at > https://github.com/sonalgoyal/hiho/

Re: any plans to deploy OSGi bundles on cluster?

2011-01-04 Thread Allen Wittenauer
On Jan 4, 2011, at 10:30 AM, Hiller, Dean (Contractor) wrote: > I guess I meant in the setting for number of tasks in child JVM before > teardown. In that case, it is nice to separate/unload my previous > classes from the child JVM which OSGi does. I was thinking we may do 10 > tasks / JVM sett

RE: any plans to deploy OSGi bundles on cluster?

2011-01-04 Thread Hiller, Dean (Contractor)
I guess I meant in the setting for number of tasks in child JVM before teardown. In that case, it is nice to separate/unload my previous classes from the child JVM which OSGi does. I was thinking we may do 10 tasks / JVM setting which I thought meant have a "Child" process run 10 tasks before shu

Re: How to split DBInputFormat?

2011-01-04 Thread Sonal Goyal
Hi Hari, I dont think DataDrivenDBInputFormat is available in 0.20.x, its only available in 0.21 versions. You can check hihoApache0.20 branch at https://github.com/sonalgoyal/hiho/ which backports the relevent db formats for Apache Hadoop 0.20 versions. Thanks and Regards, Sonal

Re: How to split DBInputFormat?

2011-01-04 Thread Joan
Thanks, I've incremented number map tasks and number of reduce tasks, Although worksI think that it's not a solution so I will try both proposals Joan 2011/1/4 Hari Sreekumar > Arvind, > > Where can I find DataDrivenInputFormat? Is it available in v0.20.2 and is > it stable? > > Thanks, > Hari

Re: When a Reduce Task starts?

2011-01-04 Thread Arun C Murthy
On Dec 23, 2010, at 9:20 PM, pig wrote: For some special reduce jobs that do not rely on the order of (key,value) pairs, the sort phase is of no use. In this situation, theoretically speaking, reduce can be started before all of the map task finished. But why hadoop doesn't support this feat