Vamshi,
What problem are you exactly trying to solve by trying to attempt
this? If you are only interested in records being streamed from one
mapper into another, why can't it be chained together? Remember that
map-only jobs do not sort their data output -- so I still see no
benefit here in consum
thank you harsh for your reply. Here what chainMapper does is, once the
first mapper finishes, then only second map starts using that file written
by first mapper. Its just like chain. But what i want is like pipelining
i.e after first map starts and before it finishes only second map has to
start
Hello,
I Have the same question at the begining of year 2010 when we start using
Hadoop...and we start using NewApi as the old ones are marked depreaceted...
And we spent lot of time using NewApi because not all the feature are not
ported in the NewApi (MultipleInputFormat for example..).
Foll
Keith
I echo Harsh's words. The old API is the stable one and the deprication
has been removed in the later releases. We do have a good number of projects in
production with old API and is definitely the recommended one.
Regards
Bejoy K S
From handheld, Please excuse typos.
-Original
Thanks! I'll look over those slides in detail, but at a first glance, they
doesn't appear to draw a final conclusion on the matter (and they're from
2009)...so, to put it bluntly, which API is the recommended usage at this
point? .20 was supposed to make it easier for maintainers of Hadoop to
Hi,
I'm trying to set a specific variable into job and I want to rehuse this
variable in other job.
The problem is when I set "context.getconfiguration().set("myvariable",xxx)
into the job, and this job finishes, then If I want to restore "myvariable"
in the other job, this variable doesn't exist