Hi guys,
Another question related to it is that when you do hadoop fs -copyFromLocal
or use
API to call fs.write(), does it write to local filesystem first before
writing to HDFS. I read and found out that it writes on local file-system
until block-size is reached and then writes on HDFS.
Wouldn't
sorry, i forgot dat, sorry, jst i am moving to a new thread.
thanks
On Thu, 26 May 2011 23:17:43 +0530, vishnu krishnan
wrote:
> thanks,
>
>
> if am not using using the map/reduce here, that just i directly sent dat
> data to the db, what will be the problems?
Look, I hate to be That Guy, especially on my first day on the list but
would you mind moving to your
thanks,
if am not using using the map/reduce here, that just i directly sent dat
data to the db, what will be the problems?
If it is just a GB then you probably don't need Hadoop, unless there is some
serious processing involved that hasn't been explained or you already have the
data on HDFS, or you happen to have a Hadoop cluster that you have access to
and the amount of data is going to grow in size. Then it could
thanku,
so just i want to take a GB of data and give to the map/reduce, then store
into the database?
--
Vishnu R Krishnan
Software Engineer
Create @ Amrita
Amritapuri
I agree. Specially people like Harsh who are always there to answer
everyone's queries !
On Wed, May 25, 2011 at 11:38 AM, Michael Giannakopoulos <
miccagi...@gmail.com> wrote:
> Thanks a lot! Your help was invaluable! Those guys like you, who answer to
> anyone are heroes! Thanks mate! Hope to t
Vishnu,
You have to have a file system that is accessible from all nodes involved to
run Hadoop Map Reduce. This could be NFS if it is a small number of nodes or
even the local file system if you are just running one node. But, with that
said Hadoop is designed to process big data GB, TB, an
am new in map reduce. one think i have to know. can i use the map reduce pgm
without any file system?
Hi, everybody.
I'm running into some difficulties getting needed libraries to map/reduce
tasks using the distributed cache.
I'm using Hadoop 0.20.2, which from what I can tell is a hard requirement
by the client, so more current versions are not really viable options.
The code I've inherited is
10 matches
Mail list logo